Greenplum Database 4.3.8.0 Release Notes
Greenplum Database 4.3.8.0 Release Notes
Rev: A01
Updated: March, 2016
Welcome to Pivotal Greenplum Database 4.3.8.0
Greenplum Database is a massively parallel processing (MPP) database server that supports next generation data warehousing and large-scale analytics processing. By automatically partitioning data and running parallel queries, it allows a cluster of servers to operate as a single database supercomputer performing tens or hundreds times faster than a traditional database. It supports SQL, MapReduce parallel processing, and data volumes ranging from hundreds of gigabytes, to hundreds of terabytes.
- Bug 1238749 - Backport rhashtable changes from upstream
- Bug 1316093 - Backport of da314c9923fe and 1f770c0a09 into RHEL-7
It is advised not to run Greenplum Database on RHEL 7 or CentOS 7 until the bugs are resolved and patched. Alternatively, a customer can contact Red Hat directly and request a fix before a kernel patch is released. Red Hat might, at its discretion, supply a kernel hotfix that includes fixes for bugs.
About Greenplum Database 4.3.8.0
Greenplum Database 4.3.8.0 is a maintenance release that resolves known issues and includes some enhancements and changes. Please refer to the following sections for more information about this release.
- Product Enhancements
- Changed Features
- New and Changed Parameters
- Supported Platforms
- Resolved Issues in Greenplum Database 4.3.8.0
- Known Issues in Greenplum Database 4.3.8.0
- Upgrading to Greenplum Database 4.3.8.0
- Greenplum Database Tools Compatibility
- Greenplum Database Extensions Compatibility
- Hadoop Distribution Compatibility
- Greenplum Database 4.3.8.0 Documentation
Product Enhancements
Greenplum Database 4.3.8.0 includes these enhancements.
External Table Support for Amazon S3 Data Source
For Greenplum Database 4.3.8.0, the CREATE EXTERNAL TABLE command supports creating readable external tables that access files from Amazon S3. For information about Amazon S3, see Amazon S3.
The s3 protocol is used in a URL that specifies the location of an Amazon Simple Storage Service (Amazon S3) bucket. Amazon S3 provides secure, durable, highly-scalable object storage. This is the s3 protocol syntax for the LOCATION clause of the CREATE EXTERNAL TABLE command.
's3://S3_endpoint/bucket_name/[folder/[folder/]...]] [config=/config_file_location]'
The s3 protocol URL specifies the AWS S3 endpoint, S3 bucket name, and optional S3 folder location.
All the files in the S3 location (S3_endpoint/bucket_name/folder_location) are used as the source for the external table and must have the same format and each file must contain complete data rows. A data row cannot be split between files. Only the TEXT and CSV formats are supported. The S3 file permissions must be Open/Download and View for the S3 user ID that is accessing the files. S3 server side encryption is not supported.
For information about the CREATE EXTERNAL TABLE command, see the Greenplum Database Reference Guide. For information about using external tables, see "Loading and Unloading" data in the Greenplum Database Administrator Guide.
gpcrondump Utility Enhancement
For Greenplum Database 4.3.8.0, the gpcrondump utility includes enhanced support for special characters. You can specify names of databases, schemes, and tables that contain these special characters.
When the name contains special characters and is specified on the command line, the name must be enclosed in double quotes ("). Double quotes are optional for names that do not contain special characters. For example, both example are valid on the command line "my#1schema".mytable. and "my#1schema"."mytable". Within the name, these special characters must be escaped with a backslash (\) : " ` $ \ .
When the name is specified in an input file, the name must not be enclosed in double quotes. Special characters do not require escaping.
For information about the utility, see the Greenplum Database Utility Guide.
gpdbrestore Utility Enhancements
These are the Greenplum Database 4.3.8.0 enhancements for the gpdbrestore utility.
- The utility supports for special characters in names of databases, schemas, and
tables. You can specify names that contain these special characters. " ' ` ~ # $ % ^ & * ( ) _ - + [ ] { } > < \ | ; : / ? and the space character.Note: The characters !, comma (,), and period (.) are not supported. Also, the tab (\t) and newline (\n) characters are not supported.
When the name contains special characters and is specified on the command line, the name must be enclosed in double quotes ("). Double quotes are optional for names that do not contain special characters. For example, both example are valid on the command line "my#1schema".mytable. and "my#1schema"."mytable". Within the name, these special characters must be escaped with a backslash (\) : " ` $ \ .
When the name is specified in an input file, the name must not be enclosed in double quotes. Special characters do not require escaping.
- The utility adds supports for the -S option to specify schema name.
All the tables that are qualified by the specified schema are restored from the
backup.
The -S option can be specified multiple times. System catalog schemas are not supported. The schema name must exist in the backup set of the database being restored. To replace the data in the schema tables with the data from backup, you can specify the --truncate option. The schema tables are truncated.
The -S option cannot be specified with the --change-schema option
- The utility performance has been improved when restoring specific append-optimized tables from a full backup (for example, with the -T or --table-file option). The utility updates statistics only append-optimized tables that are restored. In previous releases, the utility updated statistics information for all append-optimized tables in the database, not just the tables that were restored.
For information about the utility, see the Greenplum Database Utility Guide.
Pivotal Query Optimizer Enhancements
These are the Greenplum Database 4.3.8.0 enhancements for Pivotal Query Optimizer.
- For Greenplum Database 4.3.8.0, Pivotal Query Optimizer can generate query plans that include bitmap index scans. In previous releases, only the legacy query optimizer supported bitmap index scans. See resolved issue 25924.
- Greenplum Database permits GPORCA to execute catalog queries that run only on the Greenplum Database master. See the server configuration parameter optimizer_enable_master_only_queries in New and Changed Parameters.
For information about Pivotal Query Optimizer, the Greenplum Database Administrator Guide. For information about the parameter, see the Greenplum Database Reference Guide.
Changed Features
Greenplum Database 4.3.8.0 includes these changes.
- For the Greenplum Database gpcrondump utility, the -x option no longer supports specifying multiple database names separated by a comma. An error is returned.
- For the Greenplum Database gpdbrestore utility, the -T option no longer supports the * wildcard character in table names to specify a set of tables in a schema. To restore all the tables in a schema, see the gpdbrestore option -S.
- The output of Greenplum Database gpcheckcat utility has been enhanced. Each line of the output from a gpcheckcat test such as the run time and errors found include the test name. In previous releases, the name of the test was listed only when the test started.
For information about the utilities, see the Greenplum Database Utility Guide.
New and Changed Parameters
For information about Greenplum Database server configuration parameters, see the Greenplum Database Reference Guide.
New Parameter
Greenplum Database 4.3.8.0 supports the server configuration parameter optimizer_enable_master_only_queries. The parameter controls how catalog queries that run only on the Greenplum Database master are executed.
optimizer_enable_master_only_queries
When GPORCA is enabled (the server configuration parameter optimizer is on), this parameter allows the Pivotal Query Optimizer to execute catalog queries that run only on the Greenplum Database master. For the default value off, only the legacy query optimizer can execute catalog queries that run only on the Greenplum Database master.
The parameter can be set for a database system, an individual database, or a session or query.
For information about GPORCA, see the Greenplum Database Administrator Guide.
Value Range | Default | Set Classifications |
---|---|---|
Boolean | off | master session reload |
Changed Parameter
In Greenplum Database 4.3.8.0, the calculation for the server configuration parameter gp_vmem_protect_limit has been improved.
gp_vmem_protect_limit
Sets the amount of memory (in number of MBs) that all postgres processes of an active segment instance can consume. If a query causes this limit to be exceeded, memory will not be allocated and the query will fail. Note that this is a local parameter and must be set for every segment in the system (primary and mirrors).
To prevent over-allocation of memory, these calculations can estimate a safe gp_vmem_protect_limit value.
gp_vmem = ((SWAP + RAM) – (7.5GB + 0.05 * RAM)) / 1.7
where SWAP is the host swap space and RAM is the RAM on the host in GB.
Next, calculate the max_acting_primary_segments. This is the maximum number of primary segments that can be running on a host when mirror segments are activated due to a failure. With mirrors arranged in a 4-host block with 8 primary segments per host, for example, a single segment host failure would activate two or three mirror segments on each remaining host in the failed host's block. The max_acting_primary_segments value for this configuration is 11 (8 primary segments plus 3 mirrors activated on failure).
This is the calculation for gp_vmem_protect_limit. The value should be converted to MB.
gp_vmem_protect_limit = gp_vmem / acting_primary_segments
For scenarios where a large number of workfiles are generated, this is the calculation for gp_vmem that accounts for the workfiles.
gp_vmem = ((SWAP + RAM) – (7.5GB + 0.05 * RAM - (300KB * total_#_workfiles))) / 1.7
For information about monitoring and managing workfile usage, see the Greenplum Database Administrator Guide.
vm.overcommit_ratio = (RAM - (0.026 * gp_vmem)) / RAM
For information about the kernel parameter, see the Greenplum Database Installation Guide.
Value Range | Default | Set Classifications |
---|---|---|
integer | 8192 | local system restart |
Downloading Greenplum Database
These are the locations of the Greenplum Database software and documentation:
- Greenplum Database 4.3.x software is available from Pivotal Network.
- Current release Greenplum Database documentation is
available from the Pivotal Documentation site.
Previous release versions of Greenplum Database documentation, as well as other Greenplum Database documents, are available from Support Zone .
Supported Platforms
Greenplum Database 4.3.8.0 runs on the following platforms:
- Red Hat Enterprise Linux 64-bit 7.x (See Supported Platform Notes. See the Warning in Welcome)
- Red Hat Enterprise Linux 64-bit 6.x
- Red Hat Enterprise Linux 64-bit 5.x
- SuSE Linux Enterprise Server 64-bit 10 SP4, 11 SP1, 11 SP2, 11 SP4
- Oracle Unbreakable Linux 64-bit 5.5
- CentOS 64-bit 7.x
- CentOS 64-bit 6.x
- CentOS 64-bit 5.x
- 8.xxx
- 7.xxx
- 6.xxx
Greenplum Database 4.3.x supports Data Domain Boost on Red Hat Enterprise Linux.
This table lists the versions of Data Domain Boost SDK and DDOS supported by Greenplum Database 4.3.x.
Greenplum Database | Data Domain Boost | DDOS |
---|---|---|
4.3.8.0 | 3.0.0.3 | 5.6 (all versions) 5.5 (all versions) 5.4 (all versions) 5.3 (all versions) |
4.3.7.3 4.3.7.2 4.3.7.1 4.3.7.0 |
3.0.0.3 | 5.6 (all versions) 5.5 (all versions) 5.4 (all versions) 5.3 (all versions) |
4.3.7.1 4.3.7.0 |
3.0.0.3 | 5.6 (all
versions) 5.5.0.x 5.4 (all versions) 5.3 (all versions) |
4.3.6.2 4.3.6.1 4.3.6.0 |
3.0.0.3 | 5.6 (all
versions) 5.5.0.x 5.4 (all versions) 5.3 (all versions) |
4.3.5.3 4.3.5.2 4.3.5.1 4.3.5.0 |
3.0.0.3 | 5.5.0.x 5.4 (all versions) 5.3 (all versions) |
4.3.4.2 4.3.4.1 4.3.4.0 |
3.0.0.3 | 5.5.0.x 5.4 (all versions) 5.3 (all versions) |
4.3.3.0 | 2.6.2.0 | 5.2, 5.3, and 5.4 |
4.3.2.0 | 2.6.2.0 | 5.2, 5.3, and 5.4 |
4.3.1.0 | 2.6.2.0 | 5.2, 5.3, and 5.4 |
4.3.0.0 | 2.4.2.2 | 5.0.1.0, 5.1, and 5.2 |
- Greenplum Database 4.3.x, all versions, is supported on DCA V3.
- Greenplum Database 4.3.x, all versions, is supported on DCA V2, and requires DCA software version 2.1.0.0 or greater due to known DCA software issues in older DCA software versions.
- Greenplum Database 4.3.x, all versions, is supported on DCA V1, and requires DCA software version 1.2.2.2 or greater due to known DCA software issues in older DCA software versions.
Pivotal recommends that user migrate to a version of IBM Cognos software that support connecting Greenplum Database with an JDBC driver.
Supported Platform Notes
The following notes describe platform support for Greenplum Database. Please send any questions or comments to Pivotal Support at https://support.pivotal.io.
- The only file system supported for running Greenplum Database is the XFS file system. All other file systems are explicitly not supported by Pivotal.
- Greenplum Database is supported on all 1U and 2U commodity servers with local storage. Special purpose hardware that is not commodity may be supported at the full discretion of Pivotal Product Management based on the general similarity of the hardware to commodity servers.
- Greenplum Database is supported on network or shared storage if the shared storage is presented as a block device to the servers running Greenplum Database and the XFS file system is mounted on the block device. Network file systems are not supported. When using network or shared storage, Greenplum Database mirroring must be used in the same way as with local storage, and no modifications may be made to the mirroring scheme or the recovery scheme of the segments. Other features of the shared storage such as de-duplication and/or replication are not directly supported by Pivotal, but may be used with support of the storage vendor as long as they do not interfere with the expected operation of Greenplum Database at the discretion of Pivotal.
- Greenplum Database is supported when running on virtualized systems, as long as the storage is presented as block devices and the XFS file system is mounted for the storage of the segment directories.
- A minimum of 10-gigabit network is required for a system configuration to be supported by Pivotal.
- Greenplum Database is supported on Amazon Web Services (AWS) servers using either Amazon instance store (Amazon uses the volume names ephemeral[0-20]) or Amazon Elastic Block Store (Amazon EBS) storage. If using Amazon EBS storage the storage should be RAID of Amazon EBS volumes and mounted with the XFS file system for it to be a supported configuration.
- For Red Hat Enterprise Linux 7.2 or CentOS 7.2, the default systemd setting RemoveIPC=yes removes IPC connections when non-system users logout. This causes the Greenplum Database utility gpinitsystem to fail with semaphore errors. To avoid this issue, see "Setting the Greenplum Recommended OS Parameters" in the Greenplum Database Installation Guide.
Resolved Issues in Greenplum Database 4.3.8.0
The table below lists issues that are now resolved in Greenplum Database 4.3.8.0
For issues resolved in prior 4.3 releases, refer to the corresponding release notes available from Pivotal Network.
Issue Number | Category | Resolved In | Description |
---|---|---|---|
26228 | Query Optimizer, Query Planner | 4.3.8.0 | Some queries generated a segmentation fault if the query was executed by the
legacy query planner and the query performed a correlated execution against external
web tables. Now, the queries returns this message that instruct the user to use
GPORCA.
Rescan is not supported for web external table: table_name. Please use 'set optimizer=on' as a work around and 'set optimizer_enable_master_only=on' if accessing catalog tables. For information about the server configuration parameter optimizer_enable_master_only_queries. See New and Changed Parameters. For information about the server configuration parameters, see the Greenplum Database Reference Guide. For information about GPORCA, see the Greenplum Database Administrator Guide. |
26224 | Catalog and Metadata, Storage: Transaction Management | 4.3.8.0 | For DDL operations, additional validations for index operations on system catalog tables have been introduced to prevent improper catalog modifications in the case of corrupted indices on catalog tables. |
26221 | Interconnect | 4.3.8.0 | For some queries, a Greenplum Database query dispatcher received an unexpected
message from a Greenplum Database query executor. This caused Greenplum Database to
log an incorrect log message that contains this
text.Motion layer node 1 cleanup This issue has been resolved. |
26201 | Management Scripts: gptransfer | 4.3.8.0 | In some cases the Greenplum Database utility gptransfer
utility failed when the --partition-transfer option was specified.
The error message did not clearly indicate the error. Now, the utility returns an error message that describes the issue. |
26198 | Dispatch | 4.3.8.0 | In some cases, Greenplum Database did not properly handle an out of memory
condition that occurred on the master. This caused a segmentation fault. This issue has been resolved. Greenplum Database reports an ERROR for these out of memory situations. |
26176 | Query Optimizer | 4.3.8.0 | For some queries that contained a computed column and the column contained a
set returning function, Pivotal Query Optimizer incorrectly considered the column as
an unused column and dropped the column during query optimization. This issue has been resolved. |
26163 | Query Execution | 4.3.8.0 | When querying tables without statistics, the legacy query optimizer triggered
on demand statistics computation. This increased the query optimization time and
could exhaust resource queues under heavy workload. This issue has been resolved. The legacy query optimizer now uses default statistics when they are missing. If the generated plan is sub-optimal, you can run ANALYZE to generate current statistics for the table. |
26158 | Dispatch | 4.3.8.0 | For some queries, a PANIC was generated on the Greenplum Database master due to the improper handling of strings when an internal connection was started. |
26087 | Dispatch | 4.3.8.0 | A Greenplum Database PANIC was generated by a query that executed a function that was run on multiple segments and the function attempted to set a Greenplum Database server configuration parameter with the SET command. |
26078 | Management Scripts: gptoolkit | 4.3.8.0 | In the Greenplum Database gp_toolkit administrative schema, the rsqwaiterscolumn of the gp_resqueue_status table returned wrong value. |
26052 | DDL and Utility Statements | 4.3.8.0 | In some situations, a running Greenplum Database query could not be terminated
with the functions pg_cancel_backend or
pg_terminate_backend. This issue has been resolved. The functions will cancel a running query. |
26045 | Backup and Restore | 4.3.8.0 | When performing a backup operation with the Greenplum Database gpcrondump utility that specified a file with the --schema-file option, the utility deleted the file after completing the backup. |
25924 | Query Execution | 4.3.8.0 | When Pivotal Query Optimizer (PQO) is enabled, a plan that generated bitmap
index scans caused a query execution crash. This issue has been resolved and plans
that contain bitmap index scan perform as expected with PQO.
Note: In previous
releases, generating plans with bitmap index scans is disabled for Pivotal Query
Optimizer.
|
25829 | Dispatch | 4.3.8.0 | If the name of a Greenplum Database contains a quote character, SQL commands that were run in the database returned an error if the commands accessed Greenplum Database segments. |
25821 | Backup and Restore | 4.3.8.0 | Greenplum Database gpcrondump utility did not back up tables
with names that are all uppercase characters. Also, see gpcrondump utility enhancements in Product Enhancements. |
25147 | Query Optimizer | 4.3.8.0 | When GPORCA executed an ALTER TABLE
command with REORGANIZE clause, if the distribution policy of the
table is being changed to random distribution (with the DISTRIBUTED
RANDOMLY clause), the generated plan did not force a reorganization of
the tuples across Greenplum Database segments. This issue has been resolved. Greenplum Database falls back to the legacy query optimizer to execute the command. |
Known Issues in Greenplum Database 4.3.8.0
This section lists the known issues in Greenplum Database 4.3.8.0. A workaround is provided where applicable.
For known issues discovered in previous 4.3.x releases, see the release notes at Pivotal Network. For known issues discovered in other previous releases, including patch releases to Greenplum Database 4.2.x, 4.1 or 4.0.x, see the corresponding release notes, available from Dell EMC Support Zone:
Issue | Category | Description |
---|---|---|
25934 25936 |
Query Optimizer Query Planner |
For queries that compare data from columns of different character types, for
example a join comparing a columns of data types CHAR(n) and
VARCHAR(m), the returned results might not be as expected
depending the padding added to the data (space characters added after the last
non-space character). For example, this comparison returns
false.
select 'A '::char(2) ='A '::text ; This
comparison returns
true.
select 'A'::char(2) ='A '::varchar(5) ; Workaround: Pivotal recommends specifying character column types to be of data type VARCHAR or TEXT so that comparisons include padding added to the data. For information about how the character data types CHAR, VARCHAR, and TEXT handle padding added to the data see the CREATE TABLE command in the Greenplum Database Reference Guide. |
25737 | Catalog and Metadata | Greenplum Database does not support the FILTER clause within aggregate expressions. |
25754 | Management Scripts: expansion | The Greenplum Database gpexpand utility fails to create an
input file for system expansion if the Greenplum Database system define different
TCP/IP port numbers on different hosts for Greenplum Database internal
communication. Workaround: Create the input file manually. |
25833 | Management Scripts: gpexpand | The Greenplum Database utility gpexpand fails when expanding a Greenplum Database system and in the system a database table column name contains a tab character. The utility does not support database names, table names, or column names that contain a tab character. |
15835 | DDL and Utility Statements | For multi-level partitioned tables that have these characteristics:
|
12019 | Management Scripts: checkperf | When the Greenplum Database gpcheckperf utility is run with
the option -f host_file and the host that is
running gpcheckperf is listed in host_file,
processes that were started gpcheckperf might not be cleaned up
after the utility completes. Workaround: Manually stop the processes that were started by gpcheckperf. |
24870 | Query Optimizer | GPORCA might terminate all sessions if a query attempts to cast to a timestamp a date with year greater than 200,000. |
23571 | Query Optimizer | For queries that contain inequality conditions such as != , < and , >, GPORCA does not consider table indexes when generating a query plan. For those queries, indexes are not used and the query might run slower than expected. |
21508 | Query Optimizer | GPORCA does not support GiST indexes. |
20241 | Query Optimizer | For partitioned tables with indexes, GPORCA does not use indexes the if a child partition is queried directly. |
20030 | Query Optimizer | GPORCA does not support partition elimination when the query contains functions that are applied to the partition key. |
20360 | Query Execution | GPORCA does not enforce different access rights in different parts of a partition table. Pivotal recommends that you set the same access privileges for the partitioned table and all its parts (child tables). |
20241 | Query Optimizer | GPORCA does not consider indices when querying parts/child tables of partitioned tables directly. |
25326 | Interconnect | Setting the Greenplum Database server configuration parameter
log_hostname to on Greenplum Database segment
hosts causes an Interconnect Error that states that the listeneraddress name or
service not known. The parameter should be set to on only on the Greenplum Database master. |
25280 | Management Scripts: gpstart/gpstop | The Greenplum Database utility gpstop, the utility returns
an error if it is run and the system environment variable LANG is
set, for example, export LANG=ja_JP.UTF-8.
Workaround:
Unset the environment variable LANG before running the
gpstop utility. For
example:
$ unset LANG |
25246 | Management Scripts: gpconfig | When you set the server configuration parameters gp_email_to
and gp_email_from with the Greenplum Database utility
gpconfig, the utility removes the single quotes from the
values. $ gpconfig -c gp_email_to -v 'test@example.com'The improperly set parameter causes Greenplum Database to fail when it is restarted.
Workaround: Enclose the value for
gp_email_to or gp_email_from with double
quotes.
$ gpconfig -c gp_email_to -v "'test@example.com'" |
25168 | Locking, Signals, Processes | When the server configuration parameter client_min_messages is
set to either set to PANIC or FATAL and a
PANIC or FATAL level message is encountered,
Greenplum Database hangs. The client_min_messages parameter should not be set a value higher than ERROR. |
24588 | Management Scripts: gpconfig | The Greenplum Database
gpconfig utility does not display the correct information for the
server configuration parameter gp_enable_gpperfmon. The parameter
displays the state of the Greenplum Command Center data collection agents
(gpperfmon). Workaround: The SQL command SHOW displays the correct gp_enable_gpperfmon value. |
24031 | gphdfs | If a readable external table is created
with FORMAT 'CSV' and uses the gphdfs protocol, reading a record
fails if the record spans multiple lines and the record is stored in multiple HDFS
blocks. Workaround: Remove line separators from within the record so that the record does not span multiple lines. |
23824 | Authentication | In some cases, LDAP client utility tools
cannot be used after running the source command: source $GPHOME/greenplum_path.sh because the LDAP libraries included with Greenplum Database are not compatible with the LDAP client utility tools that are installed with operating system. Workaround: The LDAP tools can be used without running the source command in the environment. |
23525 | Query Planner | Some SQL queries that contain sub-selects
fail with this error. ERROR: Failed to locate datatype for paramid 0 |
23366 | Resource Management | In Greenplum Database 4.2.7.0 and later, the priority of some running queries, cannot be dynamically adjusted with the gp_adjust_priority() function. The attempt to execute this request might silently fail. The return value of the gp_adjust_priority() call indicates success or failure. If 1 is returned, the request was not successfully executed. If a number greater than 1 is returned, the request was successful. If the request fails, the priority of all running queries are unchanged, they remain as they were before the gp_adjust_priority() call. |
23492 | Backup and Restore, | A backup from a Greenplum Database 4.3.x system that is created with a Greenplum Database back up utility, for example gpcrondump, cannot be restored to a Greenplum Database 4.2.x system with the psql utility or the corresponding restore utility, for example gpdbrestore. |
23521 | Client Access Methods and Tools | Hadoop YARN based on Hadoop 2.2 or later
does not work with Greenplum Database. Workaround: For Hadoop distributions based on Hadoop 2.2 or later that are supported by Greenplum Database, the classpath environment variable and other directory paths defined in $GPHOME/lib/hadoop/hadoop_env.sh must be to be modified so that the paths point to the appropriate JAR files. |
20453 | Query Planner | For SQL queries of either of the
following
forms:SELECT columns FROM table WHERE table.column NOT IN subquery; SELECT columns FROM table WHERE table.column = ALL subquery; tuples
that satisfy both of the following conditions are not included in the result
set:
|
21838 | Backup and Restore | When restoring sets of tables with the
Greenplum Database utility gpdbrestore, the table schemas must be defined in the
database. If a table’s schema is not defined in the database, the table is not
restored. When performing a full restore, the database schemas are created when the
tables are restored. Workaround: Before restoring a set of tables, create the schemas for the tables in the database. |
21129 | DDL and Utility Statements | SSL is only supported on the master host. It is not supported on segment hosts. |
20822 | Backup and Restore | Special characters such as !, $, #, and @ cannot be used in the password for the Data Domain Boost user when specifying the Data Domain Boost credentials with the gpcrondump options --ddboost-host and --ddboost-user. |
18247 | DDL and Utility Statements |
TRUNCATE command does
not remove rows from a sub-table of a partitioned table. If you specify a sub-table
of a partitioned table with the TRUNCATE command, the command does
not remove rows from the sub-table and its child tables. Workaround: Use the ALTER TABLE command with the TRUNCATE PARTITION clause to remove rows from the sub-table and its child tables. |
19705 | Loaders: gpload |
gpload fails on Windows
XP with Python 2.6. Workaround: Install Python 2.5 on the system where gpload is installed. |
19493 19464 19426 |
Backup and Restore | The gpcrondump and
gpdbrestore utilities do not handle errors returned by DD Boost
or Data Domain correctly. These are two examples:
Workaround: The errors are logged in the master and segment server backup or restore status and report files. Scan the status and report files to check for error messages. |
15692 17192 |
Backup and Restore | Greenplum Database’s implementation of
RSA lock box for Data Domain Boost changes backup and restore requirements for
customers running SUSE. The current implementation of the RSA lock box for Data Domain Boost login credential encryption only supports customers running on Red Hat Enterprise Linux. Workaround: If you run Greenplum Database on SUSE, use NFS as your backup solution. See the Greenplum Database Administrator Guide for information on setting up a NFS backup. |
18850 | Backup and Restore | Data Domain Boost credentials cannot be
set up in some environments due to the absence of certain libraries (for example,
libstdc++) expected to reside on
the platform. Workaround: Install the missing libraries manually on the system. |
18851 | Backup and Restore | When performing a data-only restore of a particular table, it is possible to introduce data into Greenplum Database that contradicts the distribution policy of that table. In such cases, subsequent queries may return unexpected and incorrect results. To avoid this scenario, we suggest you carefully consider the table schema when performing a restore. |
18713 | Catalog and Metadata | Drop language plpgsql cascade results in
a loss of gp_toolkit
functionality. Workaround: Reinstall gp_toolkit. |
18710 | Management Scripts Suite | Greenplum Management utilities cannot
parse IPv6 IP addresses. Workaround: Always specify IPv6 hostnames rather than IP addresses |
18703 | Loaders | The bytenum field (byte offset in the load file where the error occurred) in the error log when using gpfdist with data in text format errors is not populated, making it difficult to find the location of an error in the source file. |
12468 | Management Scripts Suite |
gpexpand --rollback fails if an error occurs during expansion such that
it leaves the database down gpstart also fails as it detects that expansion is in progress and suggests to run gpexpand --rollback which will not work because the database is down. Workaround: Run gpstart -m to start the master and then run rollback. |
18785 | Loaders | Running gpload with the --ssl option and the relative path of the
source file results in an error that states the source file is
missing. Workaround: Provide the full path in the yaml file or add the loaded data file to the certificate folder. |
18414 | Loaders | Unable to define external tables with fixed width format and empty line delimiter when file size is larger than gpfdist chunk (by default, 32K). |
17285 | Backup and Restore | NFS backup with gpcrondump -c can fail. In circumstances where you haven't backed up to a local disk before, backups to NFS using gpcrondump with the -c option can fail. On fresh systems where a backup has not been previously invoked there are no dump files to cleanup and the -c flag will have no effect. Workaround: Do not run gpcrondump with the -c option the first time a backup is invoked from a system. |
17837 | Upgrade/ Downgrade | Major version upgrades internally depend
on the gp_toolkit system schema.
The alteration or absence of this schema may cause upgrades to error out during
preliminary checks. Workaround: To enable the upgrade process to proceed, you need to reinstall the gp_toolkit schema in all affected databases by applying the SQL file found here: $GPHOME/share/postgresql/gp_toolkit.sql. |
17513 | Management Scripts Suite | Running more than one gpfilespace command concurrently with
itself to move either temporary files (--movetempfilespace) or transaction files (--movetransfilespace) to a new filespace
can in some circumstances cause OID inconsistencies. Workaround: Do not run more than one gpfilespace command concurrently with itself. If an OID inconsistency is introduced gpfilespace --movetempfilespace or gpfilespace --movetransfilespace can be used to revert to the default filespace. |
17780 | DDL/DML: Partitioning |
ALTER TABLE ADD PARTITION inheritance issue When performing an ALTER TABLE ADD PARTITION operation, the resulting parts may not correctly inherit the storage properties of the parent table in cases such as adding a default partition or more complex subpartitioning. This issue can be avoided by explicitly dictating the storage properties during the ADD PARTITION invocation. For leaf partitions that are already afflicted, the issue can be rectified through use of EXCHANGE PARTITION. |
17795 | Management Scripts Suite | Under some circumstances, gppkg on SUSE is unable to correctly
interpret error messages returned by rpm. On SUSE, gppkg is unable to operate correctly under circumstances that require a non-trivial interpretation of underlying rpm commands. This includes scenarios that result from overlapping packages, partial installs, and partial uninstalls. |
17604 | Security | A Red Hat Enterprise Linux (RHEL) 6.x
security configuration file limits the number of processes that can run on
gpadmin. RHEL 6.x contains a security file (/etc/security/limits.d/90-nproc.conf) that limits available processes running on gpadmin to 1064. Workaround: Remove this file or increase the processes to 131072. |
17334 | Management Scripts Suite | You may see warning messages that
interfere with the operation of management scripts when logging in. Greenplum recommends that you edit the /etc/motd file and add the warning message to it. This will send the messages to are redirected to stdout and not stderr. You must encode these warning messages in UTF-8 format. |
17221 | Resource Management | Resource queue deadlocks may be encountered if a cursor is associated with a query invoking a function within another function. |
17113 | Management Scripts Suite | Filespaces are inconsistent when the
Greenplum database is down. Filespaces become inconsistent in case of a network failure. Greenplum recommends that processes such as moving a filespace be done in an environment with an uninterrupted power supply. |
17189 | Loaders: gpfdist |
gpfdist shows the error “Address already in use” after successfully
binding to socket IPv6. Greenplum supports IPv4 and IPv6. However, gpfdist fails to bind to socket IPv4, and shows the message “Address already in use”, but binds successfully to socket IPv6. |
16064 | Backup and Restore | Restoring a compressed dump with the
--ddboost option displays incorrect dump parameter
information. When using gpdbrestore --ddboost to restore a compressed dump, the restore parameters incorrectly show “Restore compressed dump = Off”. This error occurs even if gpdbrestore passes the --gp-c option to use gunzip for in-line de-compression. |
15899 | Backup and Restore | When running gpdbrestore with the list (-L) option, external tables do not appear; this has no functional impact on the restore job. |
Upgrading to Greenplum Database 4.3.8.0
The upgrade path supported for this release is Greenplum Database 4.2.x.x to Greenplum Database 4.3.8.0. The minimum recommended upgrade path for this release is from Greenplum Database version 4.2.x.x. If you have an earlier major version of the database, you must first upgrade to version 4.2.x.x.
Prerequisites
Before starting the upgrade process, Pivotal recommends performing the following checks.
- Verify the health of the Greenplum Database host hardware, and that you verify that the hosts meet the requirements for running Greenplum Database. The Greenplum Database gpcheckperf utility can assist you in confirming the host requirements.
- If upgrading from Greenplum Database 4.2.x.x, Pivotal recommends running the
gpcheckcat utility to check for Greenplum Database catalog
inconsistencies. Note: If you need to run the gpcheckcat utility, Pivotal recommends running it a few weeks before the upgrade and that you run gpcheckcat during a maintenance period. If necessary, you can resolve any issues found by the utility before the scheduled upgrade.
The utility is in $GPHOME/bin/lib. Pivotal recommends that Greenplum Database be in restricted mode when you run gpcheckcat utility. See the Greenplum Database Utility Guide for information about the gpcheckcat utility.
If gpcheckcat reports catalog inconsistencies, you can run gpcheckcat with the -g option to generate SQL scripts to fix the inconsistencies.
After you run the SQL scripts, run gpcheckcat again. You might need to repeat the process of running gpcheckcat and creating SQL scripts to ensure that there are no inconsistencies. Pivotal recommends that the SQL scripts generated by gpcheckcat be run on a quiescent system. The utility might report false alerts if there is activity on the system.
Important: If the gpcheckcat utility reports errors, but does not generate a SQL script to fix the errors, contact Pivotal support. Information for contacting Pivotal Support is at https://support.pivotal.io. - During the migration process from Greenplum Database 4.2.x.x, a backup is made of some files and directories in $MASTER_DATA_DIRECTORY. Pivotal recommends that files and directories that are not used by Greenplum Database be backed up, if necessary, and removed from the $MASTER_DATA_DIRECTORY before migration. For information about the Greenplum Database migration utilities, see the Greenplum Database Utility Guide.
For detailed upgrade procedures and information, see the following sections:
- Upgrading from 4.3.x to 4.3.8.0
- Upgrading from 4.3.x to 4.3.8.0 on Pivotal DCA Systems
- Upgrading from 4.2.x.x to 4.3.8.0
- For Users Running Greenplum Database 4.1.x.x
- For Users Running Greenplum Database 4.0.x.x
- For Users Running Greenplum Database 3.3.x.x
- Migrating a Greenplum Database That Contains Append-Only Tables
If you are utilizing Data Domain Boost, you have to re-enter your DD Boost credentials after upgrading from Greenplum Database 4.2.x.x to 4.3.x.x as follows:
gpcrondump --ddboost-host ddboost_hostname --ddboost-user ddboost_user --ddboost-backupdir backup_directory
Upgrading from 4.3.x to 4.3.8.0
An upgrade from 4.3.x to 4.3.8.0 involves stopping Greenplum Database, updating the Greenplum Database software binaries, upgrading and restarting Greenplum Database. If you are using Greenplum Extension packages, you must install and use Greenplum Database 4.3.5.0 or later extension packages. If you are using custom modules with the extensions, you must also use modules that were built for use with Greenplum Database 4.3.5.0 or later.
For information about the utility, see fix_ao_upgrade.py Utility.
For information about the script, see fix_visimap_owner.sql Script.
- Log in to your Greenplum Database master host as the
Greenplum administrative user:
$ su - gpadmin
- Uninstall the Greenplum Database gNet extension package if it is installed.
The gNet extension package contains the software for the gphdfs protocol. For Greenplum Database 4.3.1 and later releases, the extension is bundled with Greenplum Database. The files for gphdfs are installed in $GPHOME/lib/hadoop.
- Perform a smart shutdown of your current Greenplum
Database 4.3.x system (there can be no active connections to the database). This example
uses the -a option to disable confirmation
prompts:
$ gpstop -a
- Run the installer for 4.3.8.0 on the Greenplum Database master host. When prompted, choose an installation location in the same base directory as your current installation. For example:
/usr/local/greenplum-db-4.3.8.0
- Edit the environment of the Greenplum Database
superuser (gpadmin) and make sure you are sourcing the greenplum_path.sh file for the new
installation. For example change the following line in .bashrc or your chosen profile
file:
source /usr/local/greenplum-db-4.3.0.0/greenplum_path.sh
to:
source /usr/local/greenplum-db-4.3.8.0/greenplum_path.sh
Or if you are sourcing a symbolic link (/usr/local/greenplum-db) in your profile files, update the link to point to the newly installed version. For example:
$ rm /usr/local/greenplum-db $ ln -s /usr/local/greenplum-db-4.3.8.0 /usr/local/greenplum-db
- Source the environment file you just edited. For
example:
$ source ~/.bashrc
- Run the gpseginstall utility to install the 4.3.8.0
binaries on all the segment hosts specified in the hostfile. For
example:
$ gpseginstall -f hostfile
- Rebuild any modules that were built against the provided C language header files for use with Greenplum Database 4.3.5.0 or later (for example, any shared library files for user-defined functions in $GPHOME/lib). See your operating system documentation and your system administrator for information about rebuilding and compiling modules such as shared libraries.
- Use the Greenplum Database gppkg utility to install Greenplum Database extensions. If you were previously using any Greenplum Database extensions such as pgcrypto, PL/R, PL/Java, PL/Perl, and PostGIS, download the corresponding packages from Pivotal Network, and install using this utility. See the Greenplum Database 4.3 Utility Guide for gppkg usage details.
- After all segment hosts have been upgraded, you can
log in as the gpadmin user and restart your Greenplum Database
system:
# su - gpadmin $ gpstart
- If you are upgrading a version of Greenplum Database between 4.3.0
and 4.3.2, check your Greenplum Database for inconsistencies due to an incorrect
conversion of 4.2.x append-only tables to 4.3.x append-optimized tables.Important: The Greenplum Database system must be started but should not be running any SQL commands while the utility is running.
- Run the fix_ao_upgrade.py utility with the option
--report. The following is an
example.
$ $GPHOME/share/postgresql/upgrade/fix_ao_upgrade.py --host=mdw --port=5432 --report
- If the utility displays a list of inconsistencies, fix them by running the
fix_ao_upgrade.py utility without the --report
option.
$ $GPHOME/share/postgresql/upgrade/fix_ao_upgrade.py --host=mdw --port=5432
- (optional) Run the fix_ao_upgrade.py utility with the option --report again. No inconsistencies should be reported.
- Run the fix_ao_upgrade.py utility with the option
--report. The following is an
example.
- For databases that contain append-optimized tables that were
created from Greenplum Database 4.2.x append-only tables, run the
fix_visimap_owner.sql script. The script resolves an issue associated
with relations associated with append-optimized tables. For example, this command runs
the script on the database
testdb.
$ psql -d testdb1 -f $GPHOME/share/postgresql/upgrade/fix_visimap_owner.sql
The script displays this prompt that allows you to display changes to the affected relations without performing the operation.
Dry run, without making any modifications (y/n)?
- Enter y to list ownership changes that would have been made. The owner of the relation is not changed.
- Enter n make the ownership changes and display the changes to relation ownership.
Note: Pivotal recommends that you run the script during low activity period. Heavy workloads do not affect database functionality but might affect performance. - If you are utilizing Data Domain Boost, you have to re-enter your DD Boost credentials
after upgrading from Greenplum Database 4.3.x to 4.3.8.0 as
follows:
gpcrondump --ddboost-host ddboost_hostname --ddboost-user ddboost_user --ddboost-backupdir backup_directory
fix_visimap_owner.sql Script
The SQL script fix_visimap_owner.sql resolves ownership issues related to visimap relations that are associated with append-optimized tables.
When upgrading from Greenplum Database 4.2.x to 4.3.x, the 4.2.x append-only tables are converted to 4.3 append-optimized tables. When upgrading from 4.2.x to Greenplum Database 4.3.x earlier than 4.3.6.0, the upgrade process incorrectly assigned the owner of visimap relations to gpadmin, not the owner of the associated append-optimized table.
If you are migrating to this release Greenplum Database from a 4.3.x release earlier than 4.3.6.0, run this SQL script as the gpadmin superuser to fix the incorrect assignment issue for a database.
$GPHOME/share/postgresql/upgrade/fix_visimap_owner.sql
When you run the script, it temporarily creates two functions that update the visimap relations ownership and displays this message that lets you perform a test run without changing ownership.
Dry run, without making any modifications (y/n)?
If you enter y, the script displays the changes that would have been made. The owner of the relation is not changed.
If you enter n, the script changes the owner of the relations and displays the changes that are made.
Before exiting, the script deletes the functions it created.
fix_ao_upgrade.py Utility
The fix_ao_upgrade.py utility checks Greenplum Database for an upgrade issue that is caused when upgrading Greenplum Database 4.2.x to a version of Greenplum Database between 4.3.0 and 4.3.2.
The upgrade process incorrectly converted append-only tables that were in the 4.2.x database to append-optimized tables during an upgrade from Greenplum Database 4.2.x to a Greenplum Database 4.3.x release prior to 4.3.2.1. The incorrect conversion causes append-optimized table inconsistencies in the upgraded Greenplum Database system.
fix_ao_upgrade.py {-h master_host | --host=master_host} {-p master_port | --port=master_port} [-u user | --user=user ] [--report] [-v | --verbose] [--help]Options
- -r | --report
- Report inconsistencies without making any changes.
- -h master_host | --host=master_host
- Greenplum Database master hostname or IP address.
- -p master_port | --port=master_port
- Greenplum Database master port.
- -u user | --user=user
- User name to connect to Greenplum Database. The user must be a Greenplum Database superuser. Default is gpadmin.
- v | --verbose
- Verbose output that includes table names.
- --help
- Show the help message and exit.
If you specify the optional --report option, the utility displays a report of inconsistencies in the Greenplum Database system. No changes to Greenplum Database system are made. If you specify the --verbose option with --report, the table names that are affected by the inconsistencies are included in the output.
Dropping Orphan Tables on Greenplum Database Segments
If you upgraded to Greenplum Database 4.3.6.0 and a user dropped a table, in some cases, the table would be dropped only on the Greenplum Database master, not on the Greenplum Database segments. This created orphan tables on Greenplum Database segments. This issue occurs only with Greenplum Database 4.3.6.0. However, the orphan tables remain in Greenplum Database after upgrading to 4.3.8.0.
$GPHOME/share/postgresql/upgrade/fix_orphan_segment_tables.py
- Checks for orphan tables on segments and generates file that contains a list of the orphan tables.
- Deletes orphan tables specified in a text file.
You run the script as a Greenplum Database administrator. The script attempts to log into Greenplum Database as user who runs the script.
$GPHOME/share/postgresql/upgrade/fix_orphan_segment_tables.py -p port
To check a single database, specify the option -d database.
The command generates a list of orphan tables in the text file orphan_tables_file_timestamp. You can review the list and, if needed, modify it.
$GPHOME/share/postgresql/upgrade/fix_orphan_segment_tables.py -p port -f orphan_tables_file_timestamp
The script connects only to the databases required to drop orphan tables.
Upgrading from 4.3.x to 4.3.8.0 on Pivotal DCA Systems
Upgrading Greenplum Database from 4.3.x to 4.3.8.0 on a Pivotal DCA system involves stopping Greenplum Database, updating the Greenplum Database software binaries, and restarting Greenplum Database. If you are using Greenplum Extension packages, you must install and use Greenplum Database 4.3.5.0 or later extension packages. If you are using custom modules with the extensions, you must also use modules that were built for use with Greenplum Database 4.3.5.0 or later.
For information about the utility, see fix_ao_upgrade.py Utility.
- Log in to your Greenplum Database master host as the Greenplum
administrative user
(gpadmin):
# su - gpadmin
- Download or copy the installer file to the Greenplum Database master host.
- Uninstall the Greenplum Database gNet extension package if it is installed. For
information about uninstalling a Greenplum Database extension package, see
gppkg in the Greenplum Database Utility Guide.
The gNet extension package contains the software for the gphdfs protocol. For Greenplum Database 4.3.1 and later releases, the extension is bundled with Greenplum Database. The files for gphdfs are installed in $GPHOME/lib/hadoop.
- Perform a smart shutdown of your current Greenplum Database 4.3.x
system (there can be no active connections to the database). This example uses the
-a option to disable confirmation
prompts:
$ gpstop -a
- As root, run the Pivotal DCA installer for 4.3.8.0 on the Greenplum Database master host and specify the file
hostfile that lists all hosts in the cluster. If necessary, copy
hostfile to the directory containing the installer before running the
installer.
This example command runs the installer for Greenplum Database 4.3.8.0 for Redhat Enterprise Linux 5.x.
# ./greenplum-db-appliance-4.3.8.0-build-1-RHEL5-x86_64.bin hostfile
The file hostfile is a text file that lists all hosts in the cluster, one host name per line.
- Install Greenplum Database extension packages. For information
about installing a Greenplum Database extension package, see gppkg in
the Greenplum Database Utility Guide.Important: Rebuild any modules that were built against the provided C language header files for use with Greenplum Database 4.3.5.0 or later (for example, any shared library files for user-defined functions in $GPHOME/lib). See your operating system documentation and your system administrator for information about rebuilding and compiling modules such as shared libraries.
- After all segment hosts have been upgraded, you can log in as the
gpadmin user and restart your Greenplum Database
system:
# su - gpadmin $ gpstart
- If you are upgrading a version of Greenplum Database between 4.3.0
and 4.3.2, check your Greenplum Database for inconsistencies due to an incorrect
conversion of 4.2.x append-only tables to 4.3.x append-optimized tables.Important: The Greenplum Database system must be started but should not be running any SQL commands while the utility is running.
- Run the fix_ao_upgrade.py utility with the option
--report. The following is an
example.
$ $GPHOME/share/postgresql/upgrade/fix_ao_upgrade.py --host=mdw --port=5432 --report
- If the utility displays a list of inconsistencies, fix them by running the
fix_ao_upgrade.py utility without the --report
option.
$ $GPHOME/share/postgresql/upgrade/fix_ao_upgrade.py --host=mdw --port=5432
- (optional) Run the fix_ao_upgrade.py utility with the option --report again. No inconsistencies should be reported.
- Run the fix_ao_upgrade.py utility with the option
--report. The following is an
example.
- If you are utilizing Data Domain Boost, you have to re-enter your
DD Boost credentials after upgrading from Greenplum Database 4.3.x to 4.3.8.0 as
follows:
gpcrondump --ddboost-host ddboost_hostname --ddboost-user ddboost_user --ddboost-backupdir backup_directory
Upgrading from 4.2.x.x to 4.3.8.0
This section describes how you can upgrade from Greenplum Database 4.2.x.x or later to Greenplum Database 4.3.8.0. For users running versions prior to 4.2.x.x of Greenplum Database, see the following:
Planning Your Upgrade
Before you begin your upgrade, make sure the master and all segments (data directories and filespace) have at least 2GB of free space.
Prior to upgrading your database, Pivotal recommends that you run a pre-upgrade check to verify your database is healthy.
You can perform a pre-upgrade check by executing the gpmigrator (_mirror) utility with the --check-only option.
For example:
source $new_gphome/greenplum_path.sh; gpmigrator_mirror --check-only $old_gphome $new_gphome
Migrating a Greenplum Database That Contains Append-Only Tables
The migration process converts append-only tables that are in a Greenplum Database to append-optimized tables. For a database that contains a large number of append-only tables, the conversion to append-optimized tables might take a considerable amount of time. Pivotal supplies a user-defined function that can help estimate the time required to migrate from Greenplum Database 4.2.x to 4.3.x. For information about the user-defined function, estimate_42_to_43_migrate_time.pdf.
Append-optimized tables are introduced in Greenplum Database 4.3.0. For information about append-optimized tables, see the release notes for Greenplum Database 4.3.0.
Upgrade Procedure
This section divides the upgrade into the following phases: pre-upgrade preparation, software installation, upgrade execution, and post-upgrade tasks.
We have also provided you with an Upgrade Checklist that summarizes this procedure.
Pre-Upgrade Preparation (on your 4.2.x system)
Perform these steps on your current 4.2.x Greenplum Database system. This procedure is performed from your Greenplum master host and should be executed by the Greenplum superuser (gpadmin).
- Log in to the Greenplum Database master as the
gpadmin
user:
# su - gpadmin
- (optional)
Vacuum all databases prior to upgrade. For
example:
$ vacuumdb database_name
- (optional)
Clean out old server log files from your master and segment data directories. For
example, to remove log files from 2011 from your segment
hosts:
$ gpssh -f seg_host_file -e 'rm /gpdata/*/gp*/pg_log/gpdb-2011-*.csv'
Running VACUUM and cleaning out old logs files is not required, but it will reduce the size of Greenplum Database files to be backed up and migrated.
- Run gpstate to check for
failed segments.
$ gpstate
- If you have failed segments, you must recover
them using gprecoverseg before you can
upgrade.
$ gprecoverseg
Note: It might be necessary to restart the database if the preferred role does not match the current role; for example, if a primary segment is acting as a mirror segment or a mirror segment is acting as a primary segment.
- Copy or preserve any additional folders or files (such as backup folders) that you have added in the Greenplum data directories or $GPHOME directory. Only files or folders strictly related to Greenplum Database operations are preserved by the migration utility.
Install the Greenplum Database 4.3 Software Binaries (non-DCA)
- Download or copy the installer file to the Greenplum Database master host.
- Unzip the installer file. For
example:
# unzip greenplum-db-4.3.8.0-PLATFORM.zip
- Launch the installer using bash. For
example:
# /bin/bash greenplum-db-4.3.8.0-PLATFORM.bin
- The installer will prompt you to accept the Greenplum Database license agreement. Type yes to accept the license agreement.
- The installer will prompt you to provide an installation path. Press ENTER to accept the default install path (for example: /usr/local/greenplum-db-4.3.8.0), or enter an absolute path to an install location. You must have write permissions to the location you specify.
- The installer installs the Greenplum Database software and creates a greenplum-db symbolic link one directory level above your version-specific Greenplum installation directory. The symbolic link is used to facilitate patch maintenance and upgrades between versions. The installed location is referred to as $GPHOME.
- Source the path file from your new 4.3.8.0 installation. This example changes to the
gpadmin user before sourcing the
file:
# su - gpadmin $ source /usr/local/greenplum-db-4.3.8.0/greenplum_path.sh
- Run the gpseginstall utility to install the 4.3.8.0
binaries on all the segment hosts specified in the hostfile. For
example:
$ gpseginstall -f hostfile
Install the Greenplum Database 4.3 Software Binaries on DCA Systems
- Download or copy the installer file to the Greenplum Database master host.
- As root, run the Pivotal DCA installer for 4.3.8.0 on the Greenplum Database master host and specify
the file hostfile that lists all hosts in the cluster. If
necessary, copy hostfile to the directory containing the installer
before running the installer.
This example command runs the installer for Greenplum Database 4.3.8.0.
# ./greenplum-db-appliance-4.3.8.0-build-1-RHEL5-x86_64.bin hostfile
The file hostfile is a text file that lists all hosts in the cluster, one host name per line.
Upgrade Execution
During upgrade, all client connections to the master will be locked out. Inform all database users of the upgrade and lockout time frame. From this point onward, users should not be allowed on the system until the upgrade is complete.
- As gpadmin, source the path
file from your old 4.2.x.x installation. For
example:
$ source /usr/local/greenplum-db-4.2.8.1/greenplum_path.sh
On a DCA system, the path to the might be similar to /usr/local/GP-4.2.8.1/greenplum_path.sh depending on the installed version.
- (optional but strongly recommended) Back up all databases in your Greenplum Database system using gpcrondump. See the Greenplum Database Administrator Guide for more information on how to do backups using gpcrondump. Make sure to secure your backup files in a location outside of your Greenplum data directories.
- If your system has a standby master host
configured, remove the standby master from your system configuration. For
example:
$ gpinitstandby -r
- Perform a clean shutdown of your current
Greenplum Database 4.2.x.x system. This example uses the -a option
to disable confirmation prompts:
$ gpstop -a
- Source the path file from your new 4.3.8.0 installation. For
example:
$ source /usr/local/greenplum-db-4.3.8.0/greenplum_path.sh
On a DCA system, the path to the file would be similar to /usr/local/GP-4.3.8.0/greenplum_path.sh.
- Update the Greenplum Database environment so it
is referencing your new 4.3.8.0 installation.
- For example, update the greenplum-db symbolic link on the
master and standby master to point to the new 4.3.8.0 installation directory.
For example (as
root):
# rm -rf /usr/local/greenplum-db # ln -s /usr/local/greenplum-db-4.3.8.0 /usr/local/greenplum-db # chown -R gpadmin /usr/local/greenplum-db
On a DCA system, the ln command would specify the install directory created by the DCA installer. For example:# ln -s /usr/local/GP-4.3.8.0 /usr/local/greenplum-db
- Using gpssh, also update
the greenplum-db symbolic link on all of your segment hosts.
For example (as
root):
# gpssh -f segment_hosts_file => rm -rf /usr/local/greenplum-db => ln -s /usr/local/greenplum-db-4.3.8.0 /usr/local/greenplum-db => chown -R gpadmin /usr/local/greenplum-db => exit
On a DCA system, the ln command would specify the install directory created by the DCA installer. For example:=> ln -s /usr/local/GP-4.3.8.0 /usr/local/greenplum-db
- For example, update the greenplum-db symbolic link on the
master and standby master to point to the new 4.3.8.0 installation directory.
For example (as
root):
- (optional but recommended) Prior to
running the migration, perform a pre-upgrade check to verify that your database is
healthy by executing the 4.3.4 version of the migration utility with the
--check-only option. The command is run as
gpadmin. This example runs the gpmigrator_mirror
utility as gpadmin:
$ gpmigrator_mirror --check-only /usr/local/greenplum-db-4.2.6.3 /usr/local/greenplum-db-4.3.8.0
On a DCA system, the old GPHOME location might be similar to /usr/local/GP-4.2.8.1 (depending on the old installed version) and the new GPHOME location would be similar to /usr/local/GP-4.3.8.0.
- As gpadmin, run the 4.3.8.0
version of the migration utility specifying your old and new GPHOME
locations. If your system has mirrors, use gpmigrator_mirror. If
your system does not have mirrors, use gpmigrator. For example on a
system with
mirrors:
$ gpmigrator_mirror /usr/local/greenplum-db-4.2.7.1 /usr/local/greenplum-db-4.3.6.2
On a DCA system, the old GPHOME location might be similar to /usr/local/GP-4.2.8.1 (depending on the old installed version) and the new GPHOME location would be similar to /usr/local/GP-4.3.8.0.
Note: If the migration does not complete successfully, contact Customer Support (see Troubleshooting a Failed Upgrade ). - The migration can take a while to complete.
After the migration utility has completed successfully, the Greenplum Database 4.3.8.0 system will be running and accepting connections.
Note: After the migration utility has completed, the resynchronization of the mirror segments with the primary segments continues. Even though the system is running, the mirrors are not active until the resynchronization is complete.
Post-Upgrade (on your 4.3.8.0 system)
- If your system had a standby master host
configured, reinitialize your standby master using gpinitstandby:
$ gpinitstandby -s standby_hostname
- If your system uses external tables with gpfdist, stop all gpfdist processes on your ETL servers and reinstall gpfdist using the compatible Greenplum Database 4.3.x Load Tools package. Application Packages are available at Pivotal Network. For information about gpfdist, see the Greenplum Database 4.3 Administrator Guide.
- Rebuild any modules that were built against the provided C language header files for use with Greenplum Database 4.3.5.0 or later. (for example, any shared library files for user-defined functions in $GPHOME/lib). See your operating system documentation and your system administrator for information about rebuilding and compiling modules such as shared libraries.
- Use the Greenplum Database gppkg utility to install Greenplum Database extensions. If you were previously using any Greenplum Database extensions such as pgcrypto, PL/R, PL/Java, PL/Perl, and PostGIS, download the corresponding packages from Pivotal Network, and install using this utility. See the Greenplum Database Utility Guide for gppkg usage details.
- If you want to utilize the Greenplum Command
Center management tool, install the latest Command Center Console and update your
environment variable to point to the latest Command Center binaries (source the
gpperfmon_path.sh file from your new installation). See the
Greenplum Command Center documentation for information about installing and
configuring Greenplum Command Center.Note: The Greenplum Command Center management tool replaces Greenplum Performance Monitor.
Command Center Console packages are available from Pivotal Network.
- (optional) Check the status
of Greenplum Database. For example, you can run the Greenplum Database utility
gpstate to display status information of a running Greenplum
Database.
$ gpstate
- Inform all database users of the completed upgrade. Tell users to update their environment to source the Greenplum Database 4.3.8.0 installation (if necessary).
Upgrade Checklist
This checklist provides a quick overview of all the steps required for an upgrade from 4.2.x.x to 4.3.8.0. Detailed upgrade instructions are provided in Upgrading from 4.2.x.x to 4.3.8.0.
Pre-Upgrade Preparation (on your current system) * 4.2.x.x system is up and available |
|
![]() |
Log in to your master host as the gpadmin user (your Greenplum superuser). |
![]() |
(Optional) Run VACUUM on all databases. |
![]() |
(Optional) Remove old server log files from pg_log in your master and segment data directories. |
![]() |
Check for and recover any failed segments (gpstate, gprecoverseg). |
![]() |
Copy or preserve any additional folders or files (such as backup folders). |
![]() |
Install the Greenplum Database 4.3 binaries on all Greenplum hosts. |
![]() |
Inform all database users of the upgrade and lockout time frame. |
Upgrade Execution * The system will be locked down to all user activity during the upgrade process |
|
![]() |
Backup your current databases. |
![]() |
Remove the standby master (gpinitstandby -r). |
![]() |
Do a clean shutdown of your current system (gpstop). |
![]() |
Update your environment to source the new Greenplum Database 4.3.x installation. |
![]() |
Run the upgrade utility (gpmigrator_mirror if you have mirrors, gpmigrator if you do not). |
![]() |
After the upgrade process finishes successfully, your 4.3.x system will be up and running. |
Post-Upgrade (on your 4.3 system) * The 4.3.x.x system is up |
|
![]() |
Reinitialize your standby master host (gpinitstandby). |
![]() |
Upgrade gpfdist on all of your ETL hosts. |
![]() |
Rebuild any custom modules against your 4.3.x installation. |
![]() |
Download and install any Greenplum Database extensions. |
![]() |
(Optional) Install the latest Greenplum Command Center and update your environment to point to the latest Command Center binaries. |
![]() |
Inform all database users of the completed upgrade. |
For Users Running Greenplum Database 4.1.x.x
Users on a release prior to 4.1.x.x cannot upgrade directly to 4.3.8.0.
- Upgrade from your current release to 4.2.x.x (follow the upgrade instructions in the latest Greenplum Database 4.2.x.x release notes available at Pivotal Documentation).
- Follow the upgrade instructions in these release notes for Upgrading from 4.2.x.x to 4.3.8.0.
For Users Running Greenplum Database 4.0.x.x
Users on a release prior to 4.1.x.x cannot upgrade directly to 4.3.8.0.
- Upgrade from your current release to 4.1.x.x (follow the upgrade instructions in the latest Greenplum Database 4.1.x.x release notes available on Dell EMC Support Zone).
- Upgrade from the current release to 4.2.x.x (follow the upgrade instructions in the latest Greenplum Database 4.2.x.x release notes available at Pivotal Documentation).
- Follow the upgrade instructions in these release notes for Upgrading from 4.2.x.x to 4.3.8.0.
For Users Running Greenplum Database 3.3.x.x
Users on a release prior to 4.0.x.x cannot upgrade directly to 4.3.8.0.
- Upgrade from your current release to the latest 4.0.x.x release (follow the upgrade instructions in the latest Greenplum Database 4.0.x.x release notes available on Dell EMC Support Zone).
- Upgrade the 4.0.x.x release to the latest 4.1.x.x release (follow the upgrade instructions in the latest Greenplum Database 4.1.x.x release notes available on Dell EMC Support Zone).
- Upgrade from the 4.1.1 release to the latest 4.2.x.x release (follow the upgrade instructions in the latest Greenplum Database 4.2.x.x release notes available at Pivotal Documentation).
- Follow the upgrade instructions in these release notes for Upgrading from 4.2.x.x to 4.3.8.0.
Troubleshooting a Failed Upgrade
If you experience issues during the migration process and have active entitlements for Greenplum Database that were purchased through Pivotal, contact Pivotal Support. Information for contacting Pivotal Support is at https://support.pivotal.io.
Be prepared to provide the following information:
- A completed Upgrade Procedure.
- Log output from gpmigrator and gpcheckcat (located in ~/gpAdminLogs)
Greenplum Database Tools Compatibility
Client Tools
Greenplum releases a number of client tool packages on various platforms that can be used to connect to Greenplum Database and the Greenplum Command Center management tool. The following table describes the compatibility of these packages with this Greenplum Database release.
Tool packages are available from Pivotal Network.
Client Package | Description of Contents | Client Version | Server Versions |
---|---|---|---|
Greenplum Clients | Greenplum Database Command-Line Interface (psql) | 4.3 | 4.3 |
Greenplum Connectivity | Standard PostgreSQL Database Drivers
(ODBC, JDBC1) PostgreSQL Client C API (libpq) |
4.3 | 4.3 |
Greenplum Loaders | Greenplum Database Parallel Data Loading Tools (gpfdist, gpload) | 4.3 | 4.3 |
Greenplum Command Center | Greenplum Database management tool. | 1.3.0.2 | 4.3 |
The Greenplum Database Client Tools, Load Tools, and Connectivity Tools are supported on the following platforms:
- AIX 5.3L (32-bit)
- AIX 5.3L and AIX 6.1 (64-bit)
- Apple OSX on Intel processors (32-bit)
- HP-UX 11i v3 (B.11.31) Intel Itanium (Client and Load Tools only)
- Red Hat Enterprise Linux i386 (RHEL 5)
- Red Hat Enterprise Linux x86_64 6.x (RHEL 6)
- Red Hat Enterprise Linux x86_64 (RHEL 5)
- SUSE Linux Enterprise Server x86_64 (SLES 10 and SLES 11)
- Solaris 10 SPARC32
- Solaris 10 SPARC64
- Solaris 10 i386
- Solaris 10 x86_64
- Windows 7 (32-bit and 64-bit)
- Windows Server 2003 R2 (32-bit and 64-bit)
- Windows Server 2008 R2 (64-bit)
- Windows XP (32-bit and 64-bit)
Greenplum Database Extensions Compatibility
Greenplum Database delivers an agile, extensible platform for in-database analytics, leveraging the system’s massively parallel architecture. Greenplum Database enables turn-key in-database analytics with Greenplum extensions.
You can download Greenplum extensions packages from Pivotal Network and install them using the Greenplum Packager Manager (gppkg). See the Greenplum Database Utility Guide for details.
Note that Greenplum Package Manager installation files for extension packages may release outside of standard Database release cycles.
The following table provides information about the compatibility of the Greenplum Database Extensions and their components with this Greenplum Database release.
Pivotal supplies separate PL/Perl extension packages for Red Hat Enterprise Linux 7.x, 6.x and 5.x. Ensure you install the correct package for your operating system.
Greenplum Database Extension | Extension Components | |
---|---|---|
Name | Version | |
PostGIS 2.0.1 for Greenplum Database 4.3.x.x | PostGIS | 2.0.3 |
Proj | 4.8.0 | |
Geos | 3.3.8 | |
PL/Java 1.3 for Greenplum Database 4.3.x.x | PL/Java | Based on 1.4.0 |
Java JDK | 1.6.0_26 Update 31 | |
PL/R 2.1 for Greenplum Database 4.3.x.x | PL/R | 8.3.0.15 |
R | 3.1.0 | |
PL/R 1.0 for Greenplum Database 4.3.x.x | PL/R | 8.3.0.12 |
R | 2.13.0 | |
PL/Perl 1.2 for Greenplum Database 4.3.x.x | PL/Perl | Based on PostgreSQL 9.1 |
Perl | 5.16.3 on RHEL 7.x 5.12.4 on RHEL 6.x 5.5.8 on RHEL 5.x, SUSE 10 |
|
PL/Perl 1.1 for Greenplum Database | PL/Perl | Based on PostgreSQL 9.1 |
Perl | 5.12.4 on RHEL 5.x, SUSE 10 | |
PL/Perl 1.0 for Greenplum Database | PL/Perl | Based on PostgreSQL 9.1 |
Perl | 5.12.4 on RHEL 5.x, SUSE 10 | |
Pgcrypto 1.2 for Greenplum Database 4.3.x.x | Pgcrypto | Based on PostgreSQL 8.3 |
MADlib 1.5 for Greenplum Database 4.3.x.x | MADlib | Based on MADlib version 1.8 |
Greenplum Database 4.3.8.0 supports these minimum Greenplum Database extensions package versions.
Greenplum Database Extension | Minimum Package Version |
---|---|
PostGIS | 2.0.1 and release gpdb4.3orca |
PL/Java | 1.3 and release gpdb4.3orca |
PL/Perl | 1.2 and release gpdb4.3orca |
PL/R | 2.1 and release gpdb4.3orca |
Pgcrypto | 1.2 and release gpdb4.3orca |
MADlib | 1.9.3 and release gpdb4.3orca |
To use extension packages with Greenplum Database 4.3.8.0, you must install and use Greenplum Database extension packages (gppkg files and contrib modules) that are built for Greenplum Database 4.3.5.0 or later. For custom modules that were used with Greenplum Database 4.3.4.x and earlier, you must rebuild any modules that were built against the provided C language header files for use with Greenplum Database 4.3.8.0.
Package File Naming Convention
For Greenplum Database 4.3, this is the package file naming format.
pkgname-ver_pvpkg-version_gpdbrel-OS-version-arch.gppkg
This example is the package name for a postGIS package.
postgis-ossv2.0.3_pv2.0.1_gpdb4.3-rhel5-x86_64.gppkg
pkgname-ver - The package name and optional version of the software that was used to create the package extension. If the package is based on open source software, the version has format ossvversion. The version is the version of the open source software that the package is based on. For the postGIS package, ossv2.0.3 specifies that the package is based on postGIS version 2.0.3.
pvpkg-version - The package version. The version of the Greenplum Database package. For the postGIS package, pv2.0.1 specifies that the Greenplum Database package version is 2.0.1.
gpdbrel-OS-version-arch - The compatible Greenplum Database release. For the postGIS package, gpdb4.3-rhel5-x86_64 specifies that package is compatible with Greenplum Database 4.3 on Red Hat Enterprise Linux version 5.x, x86 64-bit architecture.
Hadoop Distribution Compatibility
This table lists the supported Hadoop distributions:
Hadoop Distribution | Version | gp_hadoop_ target_version |
---|---|---|
Pivotal HD | Pivotal HD 3.0, 3.0.1 | gphd-3.0 |
Pivotal HD 2.0, 2.1 Pivotal HD 1.0 1 |
gphd-2.0 | |
Greenplum HD | Greenplum HD 1.2 | gphd-1.2 |
Greenplum HD 1.1 | gphd-1.1 (default) | |
Cloudera | CDH 5.2, 5.3, 5.4.x, 5.5.x | cdh5 |
CDH 5.0, 5.1 | cdh4.1 | |
CDH 4.1 2 - CDH 4.7 | cdh4.1 | |
Hortonworks Data Platform | HDP 2.1, 2.2, 2.3 | hdp2 |
MapR 3 | MapR 4.x | gpmr-1.2 |
MapR 1.x, 2.x, 3.x | gpmr-1.0 | |
Apache Hadoop | 2.x | hadoop2 |
- Pivotal HD 1.0 is a distribution of Hadoop 2.0
- For CDH 4.1, only CDH4 with MRv1 is supported
- MapR requires the MapR client
Greenplum Database 4.3.8.0 Documentation
For the latest Greenplum Database documentation go to Pivotal Documentation. Greenplum documentation is provided in HTML and PDF formats.
Title | Revision |
---|---|
Greenplum Database 4.3.8.0 Release Notes | A01 |
Greenplum Database 4.3 Installation Guide | A13 |
Greenplum Database 4.3 Administrator Guide | A15 |
Greenplum Database 4.3 Reference Guide | A16 |
Greenplum Database 4.3 Utility Guide | A17 |
Greenplum Database 4.3 Client Tools for UNIX | A07 |
Greenplum Database 4.3 Client Tools for Windows | A05 |
Greenplum Database 4.3 Connectivity Tools for UNIX | A06 |
Greenplum Database 4.3 Connectivity Tools for Windows | A05 |
Greenplum Database 4.3 Load Tools for UNIX | A08 |
Greenplum Database 4.3 Load Tools for Windows | A08 |
Greenplum Command Center 2.0 Administrator Guide
Greenplum Workload Manager 1.0 User Guide |
A01 A01 |