123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650 |
- <?xml version="1.0" encoding="UTF-8"?>
- <!DOCTYPE book PUBLIC "-//OASIS//DTD DocBook XML V4.5//EN"
- "http://www.oasis-open.org/docbook/xml/4.5/docbookx.dtd">
- <book xml:base="../">
- <chapter id="Usin-HPCC-ConfigMgr">
- <title>Using Configuration Manager</title>
- <para id="cfgmgr_introP0">Configuration Manager is the utility with which
- we configure the HPCC platform. The HPCC platform's configuration is
- stored in an XML file named <emphasis
- role="bold">environment.xml</emphasis>. Once you generate an environment
- (xml) file, it gets saved into a source directory (default is <emphasis
- role="bold">/etc/HPCCSystems/source</emphasis>). You then need to stop the
- system to copy it into the active HPCC directory, then distribute it into
- place on to each node and restart the HPCC system. At no time during
- configuration do you work on the live environment file.</para>
- <para id="cfgmgr_introP1">When you install the HPCC system package, a
- default single-node environment.xml file is generated. After that, you can
- use the Configuration Manager to modify it and/or create a different
- environment file to configure components, or add nodes. There is a
- Configuration Manager wizard to help create an environment file. Give any
- environment file you create a descriptive name that would indicate what it
- is for in the source. For example, you might create an environment without
- a Roxie, you could call that file
- <emphasis>environmentNoRoxie.xml</emphasis>.</para>
- <para id="cfgmgr_p1b">You would then copy the new configuration file you
- generate from the source directory to the <emphasis
- role="bold">/etc/HPCCSystems</emphasis> directory. Rename the file to
- environment.xml, and restart the system in order to reconfigure your
- system.</para>
- <para id="cfgmgr_introP2">Configuration Manager also offers an <emphasis
- role="bold">Advanced View</emphasis> which allows more granularity for you
- to add instances of components or change the default settings of
- components for more advanced users. Even if you plan to use the Advanced
- View, it is a good idea to start with a wizard generated configuration
- file and use Advanced View to edit it.</para>
- <para id="cfgmgr_introP3">More information and specific details for each
- Configuration Manager component and attributes of those components is
- detailed in <emphasis>Using Configuration Manager</emphasis>.</para>
- <para>The following sections will provide the details for configuring an
- HPCC environment using the Configuration Manager.</para>
- <sect1 id="configuring-a-multi-node-system">
- <title>Running the Configuration Manager</title>
- <para>This section will guide you through configuring an HPCC
- environment using the Configuration Manager.</para>
- <para>The HPCC package should already be installed on ALL nodes.</para>
- <para>You can use any tool or shell script you choose.</para>
- <orderedlist>
- <listitem>
- <para>SSH to a node in your environment and login as a user with
- sudo privileges. We would suggest that it would be the first node,
- and that it is a support node, however that is up to your
- discretion.</para>
- </listitem>
- <listitem>
- <para>Start the Configuration Manager service on the node (again we
- would suggest that it should be on a support node, and further that
- you use the same node to start the Configuration Manager every time,
- but this is also entirely up to you).</para>
- <programlisting>sudo /opt/HPCCSystems/sbin/configmgr</programlisting>
- <para><graphic
- fileref="../../images/gs_img_configmgrStart.jpg" /></para>
- </listitem>
- <listitem>
- <para>Using a Web browser, go to the Configuration Manager's
- interface:</para>
- <programlisting>http://<<emphasis>ip of installed system</emphasis>>:8015</programlisting>
- <para>The Configuration Manager startup wizard displays.</para>
- </listitem>
- </orderedlist>
- <para>There are different ways to configure your HPCC system. You can
- use the <emphasis role="bold">Generate environment wizard</emphasis> and
- use that environment or experienced users can then use the <emphasis
- role="bold">Advanced View</emphasis> for more specific customization.
- There is also the option of using <emphasis role="bold">Create blank
- environment</emphasis> to generate an empty environment that you could
- then go in and add only the components you would want.</para>
- <sect2 id="Env_Wizard">
- <title>Environment Wizard</title>
- <orderedlist>
- <listitem>
- <?dbfo keep-together="always"?>
- <para>To use the wizard select the <emphasis role="bold">Generate
- new environment using wizard</emphasis> button.</para>
- <para><graphic fileref="../../images/GS_ConfigMgrWizStart.jpg"
- vendor="configmgrSS" /></para>
- </listitem>
- <listitem>
- <para>Provide a name for the environment file.</para>
- <para>This will then be the name of the configuration XML file.
- For example, we will name our environment
- <emphasis>NewEnvironment</emphasis> and this will produce a
- configuration XML file named
- <emphasis>NewEnvironment.xml</emphasis> that we will
- use.<emphasis> </emphasis></para>
- </listitem>
- <listitem>
- <para>Press the Next button.</para>
- <para>Next you will need to define the IP addresses that your HPCC
- system will be using.</para>
- </listitem>
- <listitem>
- <?dbfo keep-together="always"?>
- <para>Enter the IP addresses or hostname(s).</para>
- <para>IP Addresses can be specified individually using semi-colon
- delimiters. You can also specify a range of IPs using a hyphen
- (for example, nnn.nnn.nnn.x-y). In the image below, we specified
- the IP addresses 10.239.219.1 through 10.239.219.100 using the
- range syntax, and also a single IP 10.239.219.111.</para>
- <para><graphic fileref="../../images/GS_ConfigMgrWiz002.jpg"
- vendor="configmgrSS" /></para>
- </listitem>
- <listitem>
- <para>Press the Next button.</para>
- <para>Now you will define how many nodes to use for the Roxie and
- Thor clusters.</para>
- </listitem>
- <listitem>
- <?dbfo keep-together="always"?>
- <para>Enter the appropriate values as indicated.</para>
- <para><graphic fileref="../../images/GS_CMWiz003.jpg"
- vendor="configmgrSS" /></para>
- <variablelist>
- <varlistentry>
- <term>Number of support nodes:</term>
- <listitem>
- <para>Specify the number of nodes to use for support
- components. The default is 1.</para>
- </listitem>
- </varlistentry>
- <varlistentry>
- <term>Number of nodes for Roxie cluster:</term>
- <listitem>
- <para>Specify the number of nodes to use for your Roxie
- cluster. Enter zero (0) if you do not want a Roxie
- cluster.</para>
- </listitem>
- </varlistentry>
- <varlistentry>
- <term>Number of slave nodes for Thor cluster</term>
- <listitem>
- <para>Specify the number of slave nodes to use in your Thor
- cluster. A Thor master node will be added automatically.
- Enter zero (0) if you do not want any Thor slaves.</para>
- </listitem>
- </varlistentry>
- <varlistentry>
- <term>Number of Thor slaves per node (default 1)</term>
- <listitem>
- <para>Specify the number of Thor slave processes to
- instantiate on each slave node. Enter zero (0) if you do not
- want a Thor cluster.</para>
- </listitem>
- </varlistentry>
- <varlistentry>
- <term>Enable Roxie on demand</term>
- <listitem>
- <para>Specify whether or not to allow queries to be run
- immediately on Roxie. This must be enabled to run the
- debugger. (Default is true)</para>
- </listitem>
- </varlistentry>
- </variablelist>
- </listitem>
- <listitem>
- <para>Press the <emphasis role="bold">Next</emphasis>
- button</para>
- <para>The wizard displays the configuration parameters.</para>
- </listitem>
- <listitem>
- <?dbfo keep-together="always"?>
- <para>Press the <emphasis role="bold">Finish</emphasis> button to
- accept these values or press the <emphasis role="bold">Advanced
- View</emphasis> button to edit in advanced mode.</para>
- <graphic fileref="../../images/GS_ConfigMgrWiz004.jpg" />
- </listitem>
- </orderedlist>
- <para>You will now be notified that you have completed the
- wizard.</para>
- <para><graphic fileref="../../images/GS_ConfigMgrWiz005.jpg" /></para>
- <para>At this point, you have created a file named NewEnvironment.xml
- in the <emphasis role="bold">/etc/HPCCSystems/source</emphasis>
- directory</para>
- <informaltable colsep="1" frame="all" rowsep="1">
- <?dbfo keep-together="always"?>
- <tgroup cols="2">
- <colspec colwidth="49.50pt" />
- <colspec />
- <tbody>
- <row>
- <entry><inlinegraphic
- fileref="../../images/OSSgr3.png" /></entry>
- <entry>Keep in mind, that your HPCC configuration may be
- different depending on your needs. For example, you may not
- need a Roxie or you may need several smaller Roxie clusters.
- In addition, in a production [Thor] system, you would ensure
- that Thor and Roxie nodes are dedicated and have no other
- processes running on them. This document is intended to show
- you how to use the configuration tools. Capacity planning and
- system design is covered in a training module.</entry>
- </row>
- </tbody>
- </tgroup>
- </informaltable>
- <sect3 id="Distrib_Conf" role="brk">
- <title>Distribute the Configuration</title>
- <orderedlist>
- <listitem>
- <?dbfo keep-together="always"?>
- <para>Stop the HPCC system.</para>
- <para>If it is running stop the HPCC system (on every node),
- using a command such as this:</para>
- <para><programlisting>sudo /sbin/service hpcc-init stop</programlisting></para>
- <variablelist>
- <varlistentry>
- <term>Note:</term>
- <listitem>
- <para>You may have a multi-node system and a custom script
- such as the one illustrated in Appendix of the <emphasis
- role="bluebold">Installing and Running the HPCC
- Platform</emphasis> document to start and stop your
- system. If that is the case please use the appropriate
- command for stopping your system on every node.</para>
- </listitem>
- </varlistentry>
- </variablelist>
- <para><informaltable colsep="1" frame="all" rowsep="1">
- <tgroup cols="2">
- <colspec colwidth="49.50pt" />
- <colspec />
- <tbody>
- <row>
- <entry><inlinegraphic
- fileref="../../images/caution.png" /></entry>
- <entry>Be sure HPCC is stopped before attempting to
- copy the environment.xml file.</entry>
- </row>
- </tbody>
- </tgroup>
- </informaltable></para>
- </listitem>
- <listitem>
- <para>Back up the original environment.xml file.</para>
- <programlisting># For example
- sudo -u hpcc cp /etc/HPCCSystems/environment.xml /etc/HPCCSystems/source/environment-date.xml </programlisting>
- <variablelist>
- <varlistentry>
- <term>Note:</term>
- <listitem>
- <para>The live environment.xml file is located in your
- <emphasis role="bold">/etc/HPCCSystems/</emphasis>
- directory. Configuration Manager works on files in
- <emphasis role="bold">/etc/HPCCSystems/source</emphasis>
- directory. You must copy from this location to make an
- environment.xml file active.</para>
- </listitem>
- </varlistentry>
- </variablelist>
- <para>You can also choose to give the environment file a more
- descriptive name, to help differentiate any differences.</para>
- <para>Having environment files under source control is a good
- way to archive your environment settings.</para>
- </listitem>
- <listitem>
- <para>Copy the new .xml file from the source directory to the
- /etc/HPCCSystems and rename the file to
- <emphasis>environment.xml</emphasis></para>
- <programlisting># for example
- sudo -u hpcc cp /etc/HPCCSystems/source/NewEnvironment.xml /etc/HPCCSystems/environment.xml</programlisting>
- <para></para>
- </listitem>
- <listitem>
- <para>Copy the <emphasis
- role="bold">/etc/HPCCSystems/environment.xml</emphasis> to the
- <emphasis role="bold">/etc/HPCCSystems/</emphasis> on to
- <emphasis>every</emphasis> node.</para>
- <para>You may want to use a script to push out the XML file to
- all nodes. See the <emphasis>Example Scripts</emphasis> section
- in the Appendix of the <emphasis role="bluebold">Installing and
- Running the HPCC Platform</emphasis> document. You can use the
- scripts as a model to create your own script to copy the
- environment.xml file out to all your nodes.</para>
- </listitem>
- <listitem>
- <para>Restart the HPCC platform on all nodes.</para>
- </listitem>
- </orderedlist>
- </sect3>
- </sect2>
- </sect1>
- </chapter>
- <chapter id="Advance-SysAdmin-Topic-Chapter">
- <title>Advanced Systems Administrator Topics</title>
- <para>This chapter contains information about certain advanced HPCC
- Systems<superscript>®</superscript> Administrators topics.</para>
- <sect1 id="Admin-System-Topic">
- <title>Admin System Topics</title>
- <para>This is a System Administrative topic designed to provide some
- insight as to an aspect of System Administration for you HPCC
- System.</para>
- </sect1>
- <sect1 id="System_sizings">
- <title>System Sizings</title>
- <para>This section provides some guidance in determining the sizing
- requirements for an initial installation of HPCC. The following are some
- suggested configuration guides that can be helpful when planning your
- system.</para>
- <sect2 id="SysAdmin_MinSuggestedHW" role="nobrk">
- <title>Minimum Suggested Hardware</title>
- <para>HPCC was designed to run on common commodity hardware, and could
- function on even lesser hardware. The following list is the suggested
- minimum hardware specifications. At the very minimum you should
- consider the following hardware components for your HPCC system. These
- guidelines were put together based on real world usage of mission
- critical (uptime) with high volume data. <informaltable border="all"
- colsep="1" rowsep="1">
- <tgroup cols="3">
- <colspec colwidth="94.50pt" />
- <colspec colwidth="84.50pt" />
- <tbody>
- <row>
- <entry><emphasis role="bold">Thor slave</emphasis></entry>
- <entry>Processor</entry>
- <entry>4 x 64-bit Intel Processor per</entry>
- </row>
- <row>
- <entry></entry>
- <entry>RAM</entry>
- <entry>8GB per daemon</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Storage</entry>
- <entry>RAID - 200MB/sec Sequential Read/Write per
- node</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Network</entry>
- <entry>1 Gb/sec bandwidth</entry>
- </row>
- <row>
- <entry><emphasis role="bold">Roxie</emphasis></entry>
- <entry>Processor</entry>
- <entry>4 x 64-bit Intel Processor</entry>
- </row>
- <row>
- <entry></entry>
- <entry>RAM</entry>
- <entry>12GB per Roxie</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Storage</entry>
- <entry>400 IOPS & 2 Volumes per (RAID optional)</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Network</entry>
- <entry>1 Gb/sec bandwidth</entry>
- </row>
- <row>
- <entry><emphasis role="bold">Dali</emphasis></entry>
- <entry>Processor</entry>
- <entry>4 x 64-bit Intel Processor each</entry>
- </row>
- <row>
- <entry></entry>
- <entry>RAM</entry>
- <entry>24GB per Dali</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Storage</entry>
- <entry>RAID 1, 5, 6, 10 Volume 200GB</entry>
- </row>
- <row>
- <entry><emphasis role="bold">Other</emphasis></entry>
- <entry>Processor</entry>
- <entry>4 x 64-bit Intel Processor</entry>
- </row>
- <row>
- <entry></entry>
- <entry>RAM</entry>
- <entry>12GB</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Storage</entry>
- <entry>RAID 1, 5, 6, 10 Volume 200GB</entry>
- </row>
- <row>
- <entry></entry>
- <entry>Network</entry>
- <entry>1 Gb/sec bandwidth</entry>
- </row>
- </tbody>
- </tgroup>
- </informaltable></para>
- </sect2>
- </sect1>
- <sect1 id="Sample_Sizings">
- <title>Sample Sizings</title>
- <para>This section illustrates sample system sizings for various work
- environments. Unlike system requirements, the following samples are
- suggestions for setting up you system for various operating
- conditions.</para>
- <sect2 id="Sample-Size-HighDataVolume">
- <title>Sample Sizing for High Data volume (Typical)</title>
- <para>The most typical scenario for HPCC is utilizing it with a high
- volume of data. This suggested sample sizing would be appropriate for
- a site with large volumes of data. A good policy is to set the Thor
- size to 4 times the source data on your HPCC. Typically, Roxie would
- be about ¼ the size of Thor. This is because the data is compressed
- and the system does not hold any transient data in Roxie. Remember
- that you do not want the number of Roxie nodes to exceed the number of
- Thor nodes.</para>
- <sect3 id="SysAdm_SmplSiz_HiDataThor">
- <title>High Data Thor sizing considerations</title>
- <para>Each Thor node can hold about 2.5 TB of data (MAX), so plan
- for the number of Thor nodes accordingly for your data.</para>
- <para>If possible, SAS drives for both Thor and Roxie as they almost
- equal to SATA drives now. If not for both, get SAS drives at least
- for your Roxie cluster.</para>
- <para>Thor replicates data and is typically configured for two
- copies.</para>
- </sect3>
- <sect3 id="SysAdm_BestPrac_HiDataRoxie">
- <title>High Data Roxie sizing considerations</title>
- <para>Roxie keeps most of its data in memory, so you should allocate
- plenty of memory for Roxie. Calculate the approximate size of your
- data, and allocate appropriately. You should either increase the
- number of nodes, or increase the amount of memory.</para>
- <para>A good practice is to allocate a Dali for every Roxie
- cluster.</para>
- <para>Roxie should have a mirror. This is useful, when you need to
- update data. You update the mirror then make that primary and bring
- the other one down. This is a good practice but not really a
- necessity except in the case of high availability.</para>
- </sect3>
- </sect2>
- <sect2 id="SysAdm_BestPrac_HevyProc_LowData">
- <title>Sample Sizing for Heavy Processing on Low Data Volume</title>
- <para>The following section provides some sample sizing for heavy
- processing with approximately the amount of data indicated.</para>
- <sect3 id="SysAdm_BestPrac_750GB">
- <title>750 GB of Raw Data</title>
- <para>Thor = 3 (slaves) + 2 (management) = 5 Nodes</para>
- <para>Roxie = 3 (agents) + 1 (Dali) = 4 Nodes (This will mean that
- the environment will be down during query deployment)</para>
- <para>Spares = 2</para>
- <para>Total = 13 nodes</para>
- </sect3>
- <sect3 id="SysAdm_BestPrac_1250GB">
- <title>1250 GB of Raw Data</title>
- <para>Thor = 6 (slaves) + 2 (management) = 8 Nodes</para>
- <para>Roxie = 4 (agents) + 1 (Dali) = 5 Nodes (This will mean that
- the environment will be down during query deployment)</para>
- <para>Spares = 2</para>
- <para>Total = 17 nodes</para>
- </sect3>
- <sect3 id="SysAdm_BestPrac_2000GB">
- <title>2000 GB of Raw Data</title>
- <para>Thor = 8 (slaves) + 3 (management) = 11 Nodes</para>
- <para>Roxie = 4 (agents) + 1 (Dali) = 5 Nodes (This will mean that
- the environment will be down during query deployment)</para>
- <para>Spares = 2</para>
- <para>Total = 20 nodes</para>
- </sect3>
- <sect3 id="SysAdm_BestPrac_3500GB">
- <title>3500 GB of Raw Data</title>
- <para>Thor = 12 (slaves) + 5 (management) = 17 Nodes</para>
- <para>Roxie = 6 (agents) + 1 (Dali) = 7 Nodes (This will mean that
- the environment will be down during query deployment)</para>
- <para>Spares = 2</para>
- <para>Total = 28 nodes</para>
- </sect3>
- </sect2>
- </sect1>
- </chapter>
- </book>
|