Informatica 10.2.1 Big Data Release Notes (10.2.1) May 2018
© Copyright Informatica LLC 1998, 2020
Contents
Technical Preview...... 2 Technical Preview for Big Data Management...... 2 Technical Preview for Big Data Streaming...... 3 Technical Preview for Enterprise Data Catalog...... 3 Technical Preview for PowerExchange Adapters...... 4 Installation and Upgrade...... 5 Informatica Upgrade Paths...... 5 Support of Informatica Big Data Products in 10.2.1...... 6 Big Data Hadoop Distribution Support...... 6 Hive Run-Time Engine ...... 8 Upgrading to New Configuration...... 9 10.2.1 Fixed Limitations and Closed Enhancements...... 9 Administrator Tool Fixed Limitations (10.2.1)...... 9 Analyst Tool Fixed Limitations (10.2.1)...... 10 Application Service Fixed Limitations (10.2.1)...... 10 Big Data Management Fixed Limitations (10.2.1)...... 10 Big Data Streaming Fixed Limitations (10.2.1)...... 14 Command Line Programs Fixed Limitations (10.2.1)...... 15 Data Transformation Fixed Limitations (10.2.1)...... 15 Domain Fixed Limitations (10.2.1)...... 16 Enterprise Data Catalog Fixed Limitations (10.2.1)...... 16 Enterprise Data Lake Fixed Limitations (10.2.1)...... 17 Exception Management Fixed Limitations (10.2.1)...... 18 Mappings and Workflows Fixed Limitations...... 18 Profiles and Scorecards Fixed Limitations (10.2.1)...... 20 Security Fixed Limitations (10.2.1)...... 21 Third-Party Fixed Limitations (10.2.1)...... 21 Transformation Fixed Limitations (10.2.1)...... 22 Transformation Language Functions Fixed Limitations (10.2.1)...... 22 10.2.1 Known Limitations...... 22
2020-06-22 1 Administrator Tool Known Limitations (10.2.1)...... 23 Analyst Tool Known Limitations (10.2.1)...... 23 Application Service Known Limitations (10.2.1)...... 23 Big Data Management Known Limitations (10.2.1)...... 25 Big Data Streaming Known Limitations (10.2.1)...... 31 Domain Known Limitations (10.2.1)...... 32 Developer Tool Known Limitations (10.2.1)...... 32 Enterprise Data Catalog Known Limitations (10.2.1)...... 34 Enterprise Data Lake Known Limitations (10.2.1)...... 37 Parameters Known Limitations (10.2.1)...... 37 Profiles and Scorecards Known Limitations (10.2.1)...... 37 Reference Data Known Limitations (10.2.1)...... 38 Rule Specification Known Limitations (10.2.1)...... 39 Security Known Limitations (10.2.1)...... 39 Third-Party Known Limitations (10.2.1)...... 40 Transformations Known Limitations (10.2.1)...... 42 Workflows Known Limitations (10.2.1)...... 43 Informatica Global Customer Support...... 43
This document contains important information about restricted functionality, known limitations, and bug fixes for Informatica 10.2.1.
Technical Preview
Technical Preview for Big Data Management
Big Data Management version 10.2.1 includes functionality that is available for technical preview. Technical preview functionality is supported but is unwarranted and is not production-ready. Informatica recommends that you use in non-production environments only. Informatica intends to include the preview functionality in an upcoming GA release for production use, but might choose not to in accordance with changing market or technical circumstances. For more information, contact Informatica Global Customer Support. Intelligent Structure Discovery Effective in version 10.2.1, the intelligent structure model is available for technical preview in Amazon S3 data objects, complex file data objects, and Microsoft Azure Blob data objects. You can read files that are parsed with an intelligent structure model when you run a mapping on the Spark engine. When you add the data object to a mapping that runs on the Spark engine, you can process any input type that the model can parse.
2 2020-06-22 Python Transformation Effective in version 10.2.1, the Python transformation is available for technical preview. Use the Python transformation to execute Python code in a mapping that runs on the Spark engine. You can use a Python transformation to implement a machine model on the data that you pass through the transformation. For example, use the Python transformation to write Python code that loads a pre- trained model. You can use the pre-trained model to classify input data or create predictions.
Technical Preview for Big Data Streaming
Big Data Streaming version 10.2.1 includes functionality that is available for technical preview. Technical preview functionality is supported but is unwarranted and is not production-ready. Informatica recommends that you use in non-production environments only. Informatica intends to include the preview functionality in an upcoming GA release for production use, but might choose not to in accordance with changing market or technical circumstances. For more information, contact Informatica Global Customer Support. Python Transformation Effective in version 10.2.1, the Python transformation is available for technical preview. Use the Python transformation to execute Python code in a mapping that runs on the Spark engine. You can use a Python transformation to implement a machine model on the data that you pass through the transformation. For example, use the Python transformation to write Python code that loads a pre- trained model. You can use the pre-trained model to classify input data or create predictions. Complex Data Type in Java Transformation Effective in version 10.2.1, when you use the Java transformation, you can use complex data types to process hierarchical data. The complex data type support in the Java transformation is available for technical preview. Map Data Type Effective in version 10.2.1, you can use map data type to generate and process map data on the Spark engine. The map data type support is available for technical preview.
Technical Preview for Enterprise Data Catalog
Enterprise Data Catalog version 10.2.1 includes functionality that is available for technical preview. Technical preview functionality is supported but is unwarranted and is not production-ready. Informatica recommends that you use in non-production environments only. Informatica intends to include the preview functionality in an upcoming GA release for production use, but might choose not to in accordance with changing market or technical circumstances. For more information, contact Informatica Global Customer Support. Import from ServiceNow Effective in version 10.2.1, Catalog Administrator now connects to ServiceNow to import connections and extract the configuration metadata into the catalog. The Import from ServiceNow feature is available for technical preview.
2020-06-22 3 Database Script Resource Type Effective in version 10.2.1, you can create a Database Script resource to extract lineage information from database scripts. The Database Script resource type is available for technical preview.
Technical Preview for PowerExchange Adapters
PowerExchange Adapters version 10.2.1 includes functionality that is available for technical preview. Technical preview functionality is supported but is unwarranted and is not production-ready. Informatica recommends that you use in non-production environments only. Informatica intends to include the preview functionality in an upcoming GA release for production use, but might choose not to in accordance with changing market or technical circumstances. For more information, contact Informatica Global Customer Support. PowerExchange for Amazon S3 Effective in version 10.2.1, PowerExchange for Amazon S3 includes the following technical preview functionality:
• You can read files that are parsed with an intelligent structure model when you run a mapping on the Spark engine. When you add the data object to a mapping that runs on the Spark engine, you can process any input type that the model can parse.
• You can use Amazon S3 sources as dynamic sources in a mapping. For more information, see the Informatica PowerExchange for Amazon S3 10.2.1 User Guide. PowerExchange for HDFS Effective in version 10.2.1, you can incorporate an intelligent structure model in a complex file data object. When you add the data object to a mapping that runs on the Spark engine, you can process any input type that the model can parse. For more information, see the Informatica PowerExchange for HDFS 10.2.1 User Guide. PowerExchange for Microsoft Azure Blob Storage Effective in version 10.2.1, PowerExchange for Microsoft Azure Blob Storage includes the following technical preview functionality:
• You can run mappings on the Spark engine.
• You can read and write .csv, Avro, and Parquet files when you run a mapping on the Spark engine and in the native environment.
• You can read and write JSON when you run a mapping on the Spark engine.
• You can read files that are parsed by an intelligent structure model when you run a mapping on the Spark engine.
• You can read a directory when you run a mapping on the Spark engine.
• You can generate or skip header rows when you run a mapping in the native environment. On the Spark engine, the header row is created by default.
• You can append an existing blob. The append operation is applicable to only to the append blob and in the native environment.
4 2020-06-22 • You can override the blob or container name. In the Blob Container Override field, specify the container name or sub-folders in the root container with the absolute path.
• You can read and write .csv files compressed in the gzip format. For more information, see the Informatica PowerExchange for Microsoft Azure Blob Storage 10.2.1 User Guide.
Installation and Upgrade
Informatica Upgrade Paths
You can directly upgrade to 10.2.1 from Informatica 10.0, 10.1, 10.1.1, and 10.2. If the product version that is currently installed cannot be upgraded to Informatica 10.2.1, you must first upgrade to a supported version. To determine the Informatica product version that is currently installed, click Help > About Informatica Administrator in the Informatica Administrator header area. Note: For information about Support EOL statements, contact Informatica Global Customer Support or see, https://network.informatica.com/docs/DOC-16182 The following table describes the Informatica product versions from which you can upgrade:
Informatica Version Upgrade Path Comments
9.0.1 9.1.0 -> 9.6.1 -> 10.2 ->10.2.1 After you upgrade to version 9.1.0, you must upgrade to version 9.6.1, then upgrade to version 10.2, and then upgrade to version 10.2.1.
9.1.0 9.6.1 -> 10.2 ->10.2.1 After you upgrade to version 9.6.1, you must upgrade to version 10.2, and then upgrade to version 10.2.1.
9.5.0 9.6.1 -> 10.2 ->10.2.1 After you upgrade to version 9.6.1, you must upgrade to version 10.2, and then upgrade to version 10.2.1.
9.5.1 9.6.1 -> 10.2 ->10.2.1 After you upgrade to version 9.6.1, you must upgrade to version 10.2, and then upgrade to version 10.2.1.
9.6.0 9.6.1 -> 10.2 ->10.2.1 After you upgrade to version 9.6.1, you must upgrade to version 10.2, and then upgrade to version 10.2.1.
9.6.1 10.2 ->10.2.1 You must first upgrade to version 10.2 and then upgrade to version 10.2.1
10.0.0 10.2.1 You can directly upgrade to version 10.2.1.
10.1.0 10.2.1 You can directly upgrade to version 10.2.1.
2020-06-22 5 Informatica Version Upgrade Path Comments
10.1.1 10.2.1 You can directly upgrade to version 10.2.1.
10.2.0 10.2.1 You can directly upgrade to version 10.2.1.
Important: Version 10.2 HotFix 1 was released after version 10.2.1. Informatica does not support the upgrade from 10.2.1 to 10.2 HotFix 1, nor does it support the upgrade from 10.2 HotFix 1 to 10.2.1. The upgrade process might succeed, but you might experience results that are not supported.
Support of Informatica Big Data Products in 10.2.1
Version 10.2.1 supports Informatica big data products, such as Big Data Management and Big Data Quality. It does not support traditional products such as PowerCenter and Informatica Data Quality. If your domain contains both big data and traditional products, you must split the domain before you perform the upgrade.
Big Data Hadoop Distribution Support
Informatica big data products support a variety of Hadoop distributions. In each release, Informatica adds, defers, and drops support for Hadoop distribution versions. Informatica might reinstate support for deferred versions in a future release. The following table lists the supported Hadoop distribution versions for Informatica 10.2.1 big data products:
Product EMR HDI CDH HDP MapR
Big Data 5.10, 5.143 3.6.x 5.111, 5.121, 2.5, 2.6 6.x MEP 5.0.x2 Management 5.13, 5.14, 5.15
Big Data 5.10, 5.143 3.6.x 5.111, 5.121, 2.5, 2.6 6.x MEP 4.0.x Streaming 5.13, 5.14, 5.15
Enterprise Data N/A 3.6.x 5.13 2.6.x N/A Catalog
Enterprise Data 5.10 3.6.x 5.13 2.6.x N/A Lake
1 Big Data Management and Big Data Streaming support for CDH 5.11 and 5.12 requires EBF-11719. See KB article 533310.
2 Big Data Management support for MapR 6.x with MEP 5.0.x requires EBF-12085. See KB article 553273.
3 Big Data Management and Big Data Streaming support for Amazon EMR 5.14 requires EBF-12444. See KB article 560632.
Note: Informatica dropped support for IBM BigInsights.
6 2020-06-22 To see a list of the latest supported versions, see the Product Availability Matrix on the Informatica Customer Portal: https://network.informatica.com/community/informatica-network/product-availability-matrices.
Big Data Management Hadoop Distributions
The following table lists the supported Hadoop distribution versions and changes in Big Data Management 10.2.1:
Hadoop Distribution Supported Distribution 10.2.1 Changes Versions
Amazon EMR 5.10, 5.14 Added support for version 5.10 and 5.14. Dropped support for version 5.8.
Azure HDInsight 3.6.x Added support for version 3.6.x. Dropped support for 3.5x.
Cloudera CDH 5.11, 5.12, 5.13, 5.14, 5.15 Added support for versions 5.13, 5.14, 5.15.
Hortonworks HDP 2.5.x, 2.6.x Added support for version 2.6.x. Dropped support for version 2.4.x.
MapR 6.x MEP 5.0.x Added support for versions 6.x MEP 5.0.x. Dropped support for versions 5.2 MEP 2.0.x, 5.2.MEP 3.0.x.
Note: Informatica dropped support for IBM BigInsights. Informatica big data products support a variety of Hadoop distributions. In each release, Informatica adds, defers, and drops support for Hadoop distribution versions. Informatica might reinstate support for deferred versions in a future release. To see a list of the latest supported versions, see the Product Availability Matrix on the Informatica network: https://network.informatica.com/community/informatica-network/product-availability-matrices
Big Data Streaming Hadoop Distributions
The following table lists the supported Hadoop distribution versions and changes in Big Data Streaming 10.2.1:
Hadoop Distribution Supported Distribution 10.2.1 Changes Versions
Amazon EMR 5.10, 5.14 Added support for 5.10, 5.14. Dropped support for version 5.4.
Azure HDInsight 3.6.x Added support for version 3.6.x.
Cloudera CDH 5.11, 5.12, 5.13, 5.14, 5.15 Added support for versions 5.13, 5.14, 5.15.
2020-06-22 7 Hadoop Distribution Supported Distribution 10.2.1 Changes Versions
Hortonworks HDP 2.5.x, 2.6.x Added support for version 2.6.x. Dropped support for version 2.4.x.
MapR 6.x MEP 4.0.x Added support for versions 6.x MEP 4.0. Dropped support for versions 5.2 MEP 2.0.x, 5.2.MEP 3.0.x.
Informatica big data products support a variety of Hadoop distributions. In each release, Informatica adds, defers, and drops support for Hadoop distribution versions. Informatica might reinstate support for deferred versions in a future release. To see a list of the latest supported versions, see the Product Availability Matrix on the Informatica network: https://network.informatica.com/community/informatica-network/product-availability-matrices.
Hive Run-Time Engine
Effective in version 10.2.1, the MapReduce mode of the Hive run-time engine is deprecated, and Informatica will drop support for it in a future release. The Tez mode remains supported. Mapping When you choose to run a mapping in the Hadoop environment, the Blaze and Spark run-time engines are selected by default. Previously, the Hive run-time engine was also selected. If you select Hive to run a mapping, the Data Integration Service will use Tez. You can use the Tez engine only on the following Hadoop distributions:
• Amazon EMR
• Azure HDInsight
• Hortonworks HDP In a future release, when Informatica drops support for MapReduce, the Data Integration Service will ignore the Hive engine selection and run the mapping on Blaze or Spark. Profiles Effective in version 10.2.1, the Hive run-time engine is deprecated, and Informatica will drop support for it in a future release. The Hive option appears as Hive (deprecated) in Informatica Analyst, Informatica Developer, and Catalog Administrator. You can still choose to run the profiles on the Hive engine. Informatica recommends that you choose the Hadoop option to run the profiles on the Blaze engine.
8 2020-06-22 Upgrading to New Configuration
After you move from a Microsoft SQL server custom schema to an SQL Server database enabled with trusted connection, the test connection fails with the following error: Login failed for user 'UserName' (PLAT-8450, 460338)
10.2.1 Fixed Limitations and Closed Enhancements
Administrator Tool Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
IDQ-5941 When you create a Content Management Service, the current Data Integration Services on the domain might not appear in the service creation wizard. The issue arise when the following conditions are true: - The domain contained a node that a user deleted before you began to create the Content Management Service. - The node that the user deleted included a Data Integration Service.
PLAT-20123 Double login appears in the Administrator tool.
PLAT-18767 The Administrator tool fails to start correctly because of a duplicate 'org.apache.commons.logging.Log' file.
PLAT-17815 After you log in to the Administrator tool, the tool requires you to log in again when you access the Domain view on the Manage tab or the Monitor view.
The following table describes closed enhancement requests:
Bug Description
PLAT-19331 The Scheduler Service now automatically adjusts to Daylight Savings Time.
PLAT-19268 You can now delete custom properties in the Administrator tool.
2020-06-22 9 Analyst Tool Fixed Limitations (10.2.1)
The following table describes fixed limitations:
Bug Description
ANT-1202 When you try to open a mapping specification, the Analyst tool displays the following error: An unknown error has occurred. Check the Analyst logs or contact your administrator.
SS-50 After you upgrade to 10.1.1 HF1, the Search service index file grows too large.
Application Service Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
SS-45 The search result do not display objects in the expected order when the search relevancy is calculated based of inverse document frequency. When the frequency of a term is low in the search index, the inverse document frequency parameter value becomes higher.
PLAT-20403 DTM threads that process data on the Data Integration Service consume high CPU.
PLAT-19208 A Data Integration Service that runs on a grid fails when the home directory is a shared location between nodes on the grid.
MRS-1577 Creating a search index consumes maximum Java heap size resulting in high CPU usage for Model repository.
MRS-1486 An error appears when you run the infacmd RestoreDomain command.
MRS-1452 The Authentication failed because the authentication token is invalid error appears frequently in the Model Repository Service log. When this error appears, you cannot perform any task in the Model repository unless you restart the Model Repository Service.
MRS-1431 When you perform a search on the Model repository content, the maximum number of objects that are returned is based on the total number of Model repository objects and not the total number of mapping objects in the repository.
BDM-11629 If the infa_rpm.tar is deleted from HDFS, the Data Integration Service incorrectly determines that the infa_rpm.tar file already exists in HDFS when you run subsequent mappings.
Big Data Management Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations.
10 2020-06-22 The following table describes fixed limitations:
Bug Description
PLAT-20370 When you run a mapping task in a workflow, the Big Data Management log collector reports that the mapping logs are not found.
PLAT-18717 The Administrator tool fails to display a cluster configuration when one of the *-site.xml configuration sets contains a dot character (.) in the name, excluding the required character for the file extension.
PLAT-14603 When you enable the Model Repository Service, the following message no longer appears in the JSF logs until you enable the debug logs: [COMPONENT_0005] Failed to load component 'PortalsProvider'.
OCON-13045 A Sqoop mapping that contains an Oracle source and a Hive target writes NULL values in Decimal columns as zeroes in the target. This issue occurs if you run the Sqoop mapping on the Spark engine.
OCON-11600 If you use Kerberos clusters, Sqoop mapping logs do not provide adequate information to troubleshoot errors.
IDQ-6183 When a mapping with a transformation that reads a reference table runs in the Hive engine, the transformation reads a null value in the reference table as "/N" and returns the string "/N" as a valid value.
BDM-8517 The infacmd ms RunMapping command does not return the job ID.
BDM-6754 When the Data Integration Service is configured to run with operating system profiles and you push the mapping to an HDInsight cluster with ADLS as storage, the mapping fails with the following error: Exception Class: [java.lang.RuntimeException] Exception Message: [java.io.IOException: No FileSystem for scheme: adl]. java.lang.RuntimeException: java.io.IOException: No FileSystem for scheme: adl
BDM-2181 The summary and detail statistics empty for mappings run on Tez. (452224)
BDM-17696 When you set the tracing level to verbose in the Lookup transformation, the Data Integration Service does not write additional details to the log.
BDM-17559 When the Hive engine calculates an expression that uses decimal data types, the return value of the expression might be incorrect if the expression contains a function that honors precision and scale when evaluating arguments such as TO_CHAR or CAST.
BDM-17146 A mapping fails on the Spark engine in an SSL-enabled cluster if the mapping contains a Hive table that uses an HBase storage handler.
BDM-17100 The output file for an AWS S3 target contains a duplicate header line when the File Merge option is selected.
BDM-16971 When you run a mapping with a logical data object or flat file sources and if the mapping contains a Sorter or Joiner transformation, an error appears and the Integration Service shuts down unexpectedly.
BDM-16960 When you run a data preview or mapping as an OS profile user under the supplementary group access list, a permission denied error occurs.
2020-06-22 11 Bug Description
BDM-16955 When you run a mapping in native mode and a Filter transformation is present after an output stream of Normalizer, the data coming out of other output streams do not have a filter and incomplete data appears in the target.
BDM-16952 When you use Lookup in a mapping, the Integration Service shuts down intermittently with an assert in lkupcmgr.cpp (line # 322).
BDM-16848 The Spark engine converts leap year dates incorrectly when the date appears in the format MM- DD-YYY. For example, the Spark engine converts 02-29-2016 to 03-01-2016.
BDM-16782 Web service requests that contain an Aggregator transformation or Rank transformation and that reuse a DTM instance return incorrect responses.
BDM-16750 The Spark engine fails to generate an execution plan to create a partitioned Hive table using a custom DDL query.
BDM-16582 When you run a mapping that uses a complex file reader on the Spark engine, a memory leak causes increasing CPU usage on the Data Integration Service.
BDM-15736 When you use the Run dialog option to run a mapping, you cannot choose a different Data Integration Service to run the mapping from.
BDM-15234 Mapping generation through SQL queries takes more than 20 minutes.
BDM-15120 The ObjectIECmdImpl.java reads the INFA_OIE_VM_ARG system property before the Xmx property.
BDM-14952 Concurrent mappings that are generated at the same millisecond have same log file names with .log and .log.1 extension.
BDM-14554 An HDFS mapping that runs in the native environment against a Hadoop cluster that uses Kerberos authentication fails when it runs for a long time.
BDM-14526 After upgrading to 10.1.1 HotFix 1, when the mapping contains multiple Lookup transformations with the Lookup cache enabled, incorrect output appears and the Integration Service sometimes shuts down unexpectedly.
BDM-14522 When you validate an constant operand to a NULL expression in the Aggregator transformation, the validation fails with an error.
BDM-14348 A mapping fails on the Spark engine if the mapping creates a table in a non-default schema.
BDM-14334 Unable to output the complete list of files extracted to local data nodes during installation.
BDM-14153 The memory used to store a long and complex Hive query for string concatenation is too high.
BDM-14141 After an upgrade, mappings with a parameterized HDFS connection fail with a connection validation error.
BDM-14131 Validating a mapping to run on the Hive or Spark engine does not resolve dynamic ports for an Update Strategy transformation.
BDM-14113 After you upgrade from version 9.6.1 HotFix 3 to 10.2, the Decimal data type truncates the trailing zeroes after the Decimal point.
12 2020-06-22 Bug Description
BDM-13669 Mappings running on the Blaze engine cause a cluster failure on a cluster with more than 120 nodes.
BDM-13570 The Blaze engine fails with exit code 134 when running a profile.
BDM-13387 When running column profiling on Blaze, the mapping fails with an error like: "Invalid mapping...
BDM-12884 You can set environment variable INFA_EXTERNAL_IMPERSONATION_PROCESS variable in the OSP-supported Informatica services where you want to use your custom impersonation process.
BDM-12369 Tez fails to connect to the Application Timeline Server when you run a job on a Hortonworks HDP cluster that is enabled with Kerberos and SSL.
BDM-12163 A mapping with an unconnected Lookup transformation fails to generate an execution plan when it runs on the Blaze engine.
BDM-12361 When you create a view for a mapping that has sources with an SQL override, the fields are misaligned.
BDM-11927 When you import a cluster configuration and connectivity between the cluster and Informatica clients stops, the error message is not informative enough.
BDM-11830 When you run jobs on a grid and the connection between the Data Integration Service and connection service fails, the jobs that are further submitted to the Data Integration Service do not get executed.
BDM-11692 When you run mapping with multiple Hive sources and use a join in a query override, the data preview on a mapping specification target fails with a Java heap space error message.
BDM-11631 Using the Run Preview function for a Data Processor transformation in a cluster configuration triggers reinstallation of RPM on the cluster.
BDM-11392 When a mapping runs on the Blaze engine, the mapping reads data from Hive sources incorrectly if the physical data object and Hive source ports are not in sync.
BDM-11375 Mappings that run on the Spark engine fail because the hadoop.rpc.protection property is not correctly imported into the cluster configuration. The following error message appears: "No common protection layer between client and server."
BDM-11172 A fetch execution plan running in the Hive environment hangs for a mapping containing a Data Processor transformation with a very large number of ports.
BDM-11111 You cannot run a mapping with Sequence Generator transformation concurrently with other mappings on Data Integration Service that runs on a grid.
BDM-11109 Sqoop mappings that read binary data from Teradata and write to HBase fail.
BDM-11106 When you run a mapping that processes more than 1 GB of data on an HDInsight cluster, the Blaze engine duplicates the number of rows in the target.
BDM-11081 The Spark and Hive engines do not delete temporary directories created to run mappings.
2020-06-22 13 Bug Description
BDM-11049 After the Data Integration Service restarts, a mapping that contains a Data Processor transformation might fail to run on the Hive engine the first time that it executes.
BDM-10924 A mapping with an SQL query defined on a Hive source and target table, or with an SQL query generated on the Hive source, fails with an error like: FAILED: ParseException line
BDM-10897 When you create or import type definition libraries using JSON objects with key names that have the dot operator (.), the following validation error occurs: Evaluation failed and was not completed. Check the Developer tool logs for details.
BDM-10878 When a mapping runs on the Blaze engine, the Hive HDFS reader improperly uses a staging directory in the source HDFS location.
BDM-10837 Mapping performance is slow when the Spark engine writes to a partitioned Hive table on Amazon S3.
BDM-10670 Memory leaks occur on the Data Integration Service machine when the Spark engine heap memory reaches 4 GB.
BDM-10566 Mappings fail when you use the impersonation user to read or write files within an HDFS encryption zone in a Hadoop cluster and operating system profiles are enabled for the Data Integration Service.
BDM-10438 The option to truncate external partitioned Hive targets is not supported in BDM 10.1.1.
BDM-10098 Mappings with a Normalizer transformation fail on the Spark engine if the value in the Occurs column is large.
Big Data Streaming Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
IIS-1567 When you import complex data type definitions for a Streaming mapping, the Import Complex Data Type Definitions dialog box does not display XML in the file format list.
IIS-1152 When you create a write data object operation for a MapR Streams data object and try to view the advanced properties tab, a null pointer exception occurs.
14 2020-06-22 Bug Description
IIS-1121 When you run applications on a MapR cluster, failover does not happen and the following error occurs: ERROR StreamsListener fs/client/marlin/cc/listener/listenerimpl.cc:699 Thread: 26778 Seek called on unsubscribed partitions
IIS-1062 When you run two concurrent Streaming mappings that read from the same Kafka broker, where one mapping contains a Union transformation and one mapping contains a Joiner transformation, the mappings do not run and a warning appears in the logs as follows: WARN ConsumerCoordinator: Auto offset commit failed for group IIS-Kafka-DStream-Consumer-1499852337694: Commit cannot be completed since the group has already rebalanced and assigned the partitions to another member.
Command Line Programs Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
PLAT-20525 Nodes might hang and jobs might fail due to high CPU usage when running jobs from the command line.
PLAT-18380 The node.log file contains excessive messages with message codes LGS_10071 and LGS_10070.
PLAT-14268 Running multiple infacmd command line programs at the same time consumes excessive memory.
The following table describes closed enhancements: Bug Description
PLAT-19545 A new field disablePaging is added to the LDAP option of the infacmd isp SetLDAPConnectivity command which allows you to disable paging.
Data Transformation Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations.
2020-06-22 15 The following table describes fixed limitations:
Bug Description
B2BT-262 A Data Processor transformation might fail when it has a mapper with a preprocessor that contains the TransformByParser component.
B2BT-261 A Hierarchical to Relational Data Processor transformation for a Library might fail unexpectedly and produce the following unrelated error: There are [X] rows in group [X] that are not in the output xml. Verify that the data is sorted correctly if the input is sorted. Verify that a primary key in the parent row is not NULL.
Domain Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
PLAT-20728 The Alert Service fails when the SMTP server is configured without authentication.
PLAT-18622 When you purge a domain that contains a large number of files that must be purged, the domain becomes unavailable due to a lack of heap memory.
PLAT-11157 When the domain fails to respond to a connection initialization message within the timeout period, the Data Transport Framework incorrectly reports an invalid error.
The following table describes closed enhancement requests:
Bug Description
PLAT-19630 When you define a domain, you can enable password complexity to validate the password strength.
PLAT-18366 Client side errors are now handled by the Node process without initiating its shutdown.
Enterprise Data Catalog Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
LDM-6588 An error appears when you run an SAP R/3 resource with the First N rows sampling option.
LDM-6555 When you run the profiling scanner on a Microsoft SQL Server data object, the Invalid utf8 data error appears.
16 2020-06-22 Bug Description
LDM-6215 When you run a Teradata resource after you choose a sampling option, the column profile results take a long time to appear. This issue occurs because Enterprise Data Catalog runs the column profile on full tables instead of a data source sample.
LDM-6109 When you run a resource with the Random sampling option, the error message SEVERE message "FAILED to fetch statistic appears in the Data Integration Service logs even when the profile run is successful.
LDM-5997 The Metadata Load Settings > Basic Profile Settings > Source Connection Name option does not display the correct tool tip.
LDM-5544 When you run a column profile on an XML data object, the data validation error appears if a column field name in the data object contains the - character.
EIC-3269 If you enable the Similarity Profile option and choose First N rows sampling option for a Microsoft SQL Server 2014 resource or Microsoft SQL Server 2016 resource that has more than 250 columns, and you run the resource on the Blaze engine, the similarity scanner fails with null pointer exception.
EIC-3152 Inconsistent profile results or value frequency results appear when you run a profile on Random N rows of a Microsoft SQL Server 2014 resource or Microsoft SQL Server 2016 resource on the Blaze engine.
EIC-2669 Similarity profile runs on all the rows in the data object even when you configure the sampling option as First N rows.
Enterprise Data Lake Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
IDL-7995 When you apply a filter containing multibyte characters in the Data Preview tab, the preview does not return any data.
IDL-6679 When you upload a delimited file, if you set the Hive Table Storage Format property for the Enterprise Data Lake Service to Parquet or ORC and select a non-default schema to publish the asset to, Enterprise Data Lake creates a table for the asset in the incorrect schema.
IDL-6556 When you publish a worksheet in which a numeric column included in an aggregation calculation is inferred as a string, Enterprise Data Lake generates an invalid mapplet.
IDL-6410 When you perform an aggregate calculation on a worksheet, the operation excludes columns that contain a single numeric value.
IDL-5531 The data preparation page for a project fails to open if the web browser you use to access Enterprise Data Lake does not have the self-signed certificates or certificate authority (CA)-signed certificates installed on the node hosting the Enterprise Data Lake Service.
IDL-5102 Data preview of a postgres table fails if the schema name consists of lower case characters.
2020-06-22 17 Bug Description
IDL-5101 Data preview of a MySQL table that includes the BIGINT UNSINGED data type fails with the following error: Failed to preview data. Possible reasons include insufficient privileges to view the data. Contact your administrator or choose a different connection.
IDL-4987 If you log out of Intelligent Data Lake and log in, the Visualization tab does not appear in the data asset views.
IDL-4973 If the system directory is encrypted, upload to the Hive database fails.
IDL-4955 In a fully profiled data asset, decimal data appears as null values after the data asset is published.
IDL-4832 If the data asset contains ultibytes and extended ASCII characters, data preview and tabular preview show incorrect data.
IDL-1716 Publication of a worksheet that includes an inner join fails if Enterprise Data Lake infers string columns in the worksheet as Integers.
BDM-105 If you create encrypted zones in the cluster and run the mappings or preview data in native mode, the 66 mapping fails.
The following table describes closed enhancement requests:
Bug Description
IDL-5514 Enterprise Data Lake supports LDAP authentication for HiveServer2.
Exception Management Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
IDQ-5701 The Analyst tool displays an error when you update a cell to a null value in a cluster step. The issue arises when the column that contains the cell uses the varchar(1) data type.
Mappings and Workflows Fixed Limitations
Review the Release Notes of previous releases for information about previous fixed limitations.
18 2020-06-22 The following table describes fixed limitations:
Bug Description
PLAT-20528 A mapping fails when the mapping is configured with parallelism set to Auto and the Data Integration Service is configured with maximum parallelism set to a value greater than 1.
OCON-12314 Concurrent dynamic mappings display an incorrect error message in the mapping log when concurrency issues occur.
OCON-12105 Dynamic mappings fail intermittently with the following error message: [com.informatica.sdk.dtm.InvalidMappingException] Exception Message: [[LDTMEXP_0029] Failed to process mapping because of following reason [null].].
MWF-1571 A workflow terminates when it tries to parse a system variable date.
MWF-1495 You can configure multiple sequence flows with the same name between two gateways in a workflow.
MWF-1493 If you run multiple concurrent workflows with Mapping tasks continuously over 24 hours in a hadoop environment, the Data Integration Service spawns an excessive number of threads and eventually stops.
MWF-1479 If you restart the workflow database, the Workflow Orchestration Service module on the Data Integration Service becomes unavailable.
MWF-1465 A workflow might fail to start when a mapping in the workflow contains a database connection that is not valid, even if the workflow does not call the database connection at run time.
MWF-1448 If a domain fails over from one gateway node to another gateway node while a workflow runs, the Data Integration Service re-runs any Mapping task that began before the domain failed. Any mapping that runs again fails during workflow recovery. The issue arises when the following conditions are true: - The workflow contains multiple parallel Mapping tasks and Command tasks between two Inclusive gateways. - The batch execution pool size specifies that the mappings in the Mapping tasks must run in two or more batches. - You manually recover the workflow or the workflow recovers automatically after the domain fails over. - You configure the nodes in a grid, and the grid Data Integration Service runs the workflow.
IDQ-5978 When an Exclusive gateway receives more than one default sequence flow from upstream gateways in a workflow, all of the default sequence flows that enter the gateway can run.
BDM-17539 When the Data Integration Service is disabled, the error message in the Developer tool incorrectly states that operating system profiles are disabled.
BDM-16777 If a master Data Integration Service loses contact with a worker Data Integration Service during a workflow run, the master service does not try to reconnect to the worker service. The master Data Integration Service defines the workflow status as failed.
2020-06-22 19 Bug Description
BDM-13914 A mapping that includes an Address Validator transformation can fail when the following conditions are true: - You set the Max Parallelism property on the mapping to Auto. - You set the Max Parallelism property on the Data Integration Service to a higher value than the Execution instances property on the transformation.
BDM-13067 A workflow fails during a Mapping task because workflow validation does not detect an unresolved parameter or connection in a mapplet that you embedded in a Match transformation.
Profiles and Scorecards Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
IDE-3157 In the Analyst tool, drilldown does not work as expected for column profile results if you configured the 'ExecutionContextOptions.drilldown_page_size custom property for Data Integration Service as n and if the number of rows returned by drilldown exceeds n.
IDE-3154 When the connection objects in the Informatica domain exceeds 500, column profiles take a long time to open in Informatica Developer.
IDE-3142 A column profile with a sampling option fails for Microsoft SQL Server data objects if the schema name contains the $ character.
IDE-3095 When an operating system profile user runs a column profile on a flat file in Informatica Analyst, the The source data object was deleted. The profile results are from the last profile run. warning message appears if you configure the Source file directory as SourceDir (Parameter) for the flat file.
IDE-3056 Column profile run fails for a logical data object when the profile name exceeds 128 characters and you run the profile on the Hive engine in Hadoop run-time environment.
IDE-3052 When an operating system profile user runs a column profile, the profile run fails if the Use the Logged in User as Hadoop Impersonation User property for operating system profile user is set to true.
IDE-3028 In Informatica Developer, an exception appears when you export the results summary for an enterprise discovery profile.
IDE-3025 You can run profiles even when you do not have the Run Profiles and Scorecards privilege.
IDE-3009 Sometimes, the profile run fails when you run the enterprise discovery profile for Microsoft SQL Server schema objects in the Analyst tool.
IDE-2936 When you add a rule the second time to the scorecard, the previous scorecard run results display 0 rows.
IDE-2859 Profiling results do not appear for some column profiles in the Analyst tool but appears in the Developer tool.
20 2020-06-22 Security Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
PLAT-19421 When you modify the LDAP synchronization schedule in the Administrator tool, the Administrator might continue to display and use the former schedule.
BDM-9945 The Data Integration Service cannot use user impersonation to read data from an HDFS file that is located in an encryption zone.
BDM-14496 The Hive engine fails to run a Hive query due to a Kerberos authentication error if the query takes a long time to run.
BDM-12237 Kerberos authentication fails when you run a mapping that reads from an HDFS source and writes to an HDFS target on the Spark engine.
BDM-11116 When a mapping runs on the Blaze engine, the mapping fails while authenticating on Hive if the following conditions are true: - Hiveserver2 is enabled with LDAP authentication. - The mapping reads from a Hive source and writes to a Hive target. The following error occurs: Peer indicated failure: Error validating the log
BDM-11092 Mappings fail when the mappings run longer than the Kerberos ticket expiration time on a Kerberos-enabled Hadoop cluster.
The following table describes closed enhancement requests:
Bug Description
PLAT-18537 You can disable LDAP paging for groups in an LDAP security domain.
PLAT-16989 The domain log contains details on user lock out operations.
Third-Party Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
BDM-17527 Mappings that run on Hadoop clusters that use Hive 1.x do not pad zeroes to match the scale of the column precision.. Reference number: HIVE-12063
2020-06-22 21 Transformation Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
OCON-13365 The REST Web Service Consumer transformation does not parse an unnamed array with a single instance of element.
IDQ-5965 If you run multiple identity match mappings concurrently through web services, the Data Transformation Manager (DTM) might fail to open the identity population data files.
IDQ-5651 If the Address Validator transformation reads a United States address that includes a character space between the digits in the house number, the mapping that contains the transformation stops unexpectedly.
Transformation Language Functions Fixed Limitations (10.2.1)
Review the Release Notes of previous releases for information about previous fixed limitations. The following table describes fixed limitations:
Bug Description
BDM-14555 The return value of an expression might be incorrect when the following conditions are true: - The expression uses an argument that is a decimal port with scale 0. - The input to the decimal port is a constant. - The expression contains a function that honors precision and scale when evaluating arguments such as TO_CHAR or CAST.
BDM-10790 NULL data is loaded to the Hive target when you use a decimal with a precision less than 29 as an argument in the function RTRIM or LTRIM.
10.2.1 Known Limitations
22 2020-06-22 Administrator Tool Known Limitations (10.2.1)
The following table describes known limitations:
Bug Description
PLAT-20807 When you try to access search from the Administrator tool or open the error codes from the logs in Informatica Knowledge Base, the following link is not accessible: https://csm.informatica.com/informaticaCSM/infa9/csm/ACProxy/oktaLogin.jsp Workaround: Use the following Network link to access search from the Administrator tool or open the error codes from the logs in Informatica Knowledge Base: https://network.informatica.com/login!input.jspa?referer=https://csm.informatica.com
OCON-1138 When you import or export data through Sqoop, the Administrator tool does not display the correct execution statistics in the Monitoring tab. Workaround: See the execution statistics in the yarn log. (452798)
Analyst Tool Known Limitations (10.2.1)
The following table describes known limitations:
Bug Description
ANT-682 When you try to delete an asset that another user changed, the Analyst tool fails to warn you that the asset is not the latest copy. (396636)
ANT-595 After you rename a domain, you cannot export a mapping specification to PowerCenter. Workaround: Use the Developer tool to export the mapping object.
ANT-45 When you try to export the mapping specification to PowerCenter Repository using pcclientsmartuser that run on a Windows network using two-factor authentication, the mapping specification export fails. Workaround: Export the mapping specification to PowerCenter Repository using INFAKRB.INFADEV.COM (460405)
Application Service Known Limitations (10.2.1)
The following table describes known limitations:
Bug Description
OCON-13454 The Developer tool fails to connect to an active Metadata Access Service process even if the Metadata Access Service is configured for high availability.
OCON-13257 When you import a complex file object, an HBase object, or a Hive object from a Hadoop cluster, the metadata import fails if the domain and the Metadata Access Service use Kerberos authentication.
MRS-1617 When you configure the monitoring Model repository on an Oracle database, an exception appears when the number of objects fetched by the monitoring Model Repository Service for the mapping job statistics exceeds 1000 objects.
2020-06-22 23 Bug Description
MRS-1585 An error appears when you enable Git version control system for a Model repository if the global repository name shares the same name with another Model repository, and if both the repositories share the local repository. Workaround: Use separate local repository for the Model Repository Services.
MRS-1571 When you enable the Model Repository Service for DEBUG level of logging, the service generates multiple ERROR level log messages.
MRS-1570 After you assign the Show Security Details privilege, you cannot view upstream dependencies for a reusable Model repository object unless you have read permission on the dependent projects.
MRS-1463 The version history for Model Repository Service objects are lost if the content of the versioned Model repository is migrated from one Perforce depot to another.
MRS-1462 You cannot restore or view the older versions of Model repository objects after you restore the contents of a Subversion repository to another Subversion repository.
BDM-4669 The Data Integration Service does not apply the cost-based optimization method when you configure the mapping to use load order constraints with the full optimizer level. (431534)
BDM-2483 The Processes tab of the Email Service includes an environment variable section even though environment variables are not supported for the Email Service. If you add an enviornment variable, the Email Service ignores it. (442102)
BDM-1828 If you run web service requests on a Data Integration Service grid and you incorrectly configure the external HTTP load balancer to use nodes with the service role only, the Data Integration Service does not redirect requests to nodes with both the service and compute roles. Some web service requests dispatched to the node with the service role only might fail. Workaround: Configure the external HTTP load balancer to use nodes with both the service and compute roles. (427052)
BDM-1798 When you run a mapping on a Data Integration Service grid configured to run jobs in separate remote processes, the Monitor tab of the Administrator tool might indefinitely list the mapping state as Running even though the infacmd command line program and the mapping log indicate that the mapping failed. (432316)
BDM-19090 The Data Integration Service hangs when the service runs concurrent SQL data service mappings.
BDM-17402 When the Data Integration Service runs on a grid, the service waits for applications to load before the service starts up.
BDM-17301 The Data Integration Service restarts if deployed applications take more than 15 minutes to load during service start-up.
BDM-17293 Application details cannot be fetched from the Data Integration Service while the Data Integration Service is deploying applications.
BDM-17055 Requests to submit mappings hang if the application that contains the mapping is loading on the Data Integration Service during service start-up.
24 2020-06-22 Bug Description
BDM-13006 Mappings run on a Data Integration Service grid might hang indefinitely when the following conditions are true: - The Data Integration Service is configured to run jobs in separate remote processes. - The Resource Manager Service becomes unavailable after the Data Integration Service has been enabled and has elected a master compute node. Workaround: Enable the Resource Manager Service to continue running the mappings. (439628)
BDM-13004 The DTM process does not create DTM log files for mappings included in workflow Mapping tasks when the following conditions are true: - The Data Integration Service is configured to run jobs in separate remote processes. - The mapping included in the workflow Mapping task uses multibyte characters. (443052)
BDM-13002 A Data Integration Service grid configured to run jobs in separate remote processes does not use a secure connection to communicate with remote DTM processes even though the domain is enabled for secure communication. (432752)
BDM-12991 In a Kerberos domain, mappings fail to run on a Data Integration Service grid configured to run jobs in separate remote processes. Workaround: Configure the Data Integration Service to run jobs in separate local processes. (435471)
BDM-12990 When you update the compute role on a node assigned to a Data Integration Service grid and then recycle the Data Integration Service, you might encounter inconsistent behavior across the Informatica client tools. For example, mappings might fail to run in the infacmd command line program but succeed in the Developer tool. Workaround: Restart the domain. (436753)
BDM-12987 The consolidated log file for a mapping might contain the incorrect DTM log file when the following conditions are true: - The Data Integration Service is configured to run jobs in separate remote processes. - The Mapping task in a workflow is configured to save the Mapping task log file by the number of Mapping task runs. Workaround: Configure the Mapping task to save the Mapping task log file by timestamp. (439632)
Big Data Management Known Limitations (10.2.1)
The following table describes known limitations:
Bug Description
OCON-9377 When you configure Sqoop and run a Teradata Parallel Transporter mapping on a Cloudera cluster to export data of the Byte or Varbyte data type to a Teradata target, the mapping fails on the Blaze engine.
OCON-9376 If you configure Sqoop to export data of the Blob or Clob data type to a Teradata target, TDCH mappings fail on the Spark engine.
OCON-9143 In the read and write operations for a complex file data object, you cannot edit the precision and scale of elements within a field that is of a complex data type. For example, if Field1 is of type array with string elements, you cannot edit the precision and scale of the string elements.
2020-06-22 25 Bug Description
OCON-9005 When you run TDCH mappings on the Hive engine to write time data to a Teradata target, the nanosecond part is truncated.
OCON-8850 If you configure Sqoop to export data of the Timestamp data type from a Hive source to a Microsoft Azure SQL Data Warehouse target, the mapping fails.
OCON-8779 If you configure Sqoop to export data of the Real data type to IBM DB2 z/OS targets, the mapping fails.
OCON-7687 When you export data through Sqoop and the columns contain mixed case characters, the mapping fails.
OCON-7669 When you configure Sqoop and OraOop, and export data to an Oracle target that contains mixed case characters in the table name, the mapping fails. Workaround: Use the generic Oracle JDBC driver to export data.
OCON-7429 When you run a Teradata Parallel Transporter mapping on a Hortonworks cluster and on the Blaze engine to write data of the Byte or Varbyte data type to a Teradata target, the data gets corrupted. This issue occurs when you use the hdp-connector-for-teradata-1.5.1.2.5.0.0-1245- distro.tar.gz JAR. Workaround: Use the hdp-connector-for-teradata-1.4.1.2.3.2.0-2950-distro.tar.gz JAR.
OCON-730 When you export data through Sqoop and there are primary key violations, the mapping fails and bad records are not written to the bad file. (456616)
OCON-7291 Mappings that read data from a Teradata source and contain the != (not equal) operator in the filter override query fail. This issue occurs if you run the Teradata Parallel Transporter mapping on a Hortonworks cluster and on the Blaze engine. Workaround: Use a native expression with the ne operator instead of the != operator.
OCON-7280 If you configure Sqoop and update the columns in the advanced SQL query, the mapping fails on the Blaze engine.
OCON-7216 If a Sqoop source or target contains a column name with double quotes, the mapping fails on the Blaze engine. However, the Blaze Job Monitor incorrectly indicates that the mapping ran successfully and that rows were written into the target.
OCON-7212 If there are unconnected ports in a target, Sqoop mappings fail on the Blaze engine. This issue occurs when you run the Sqoop mapping on any cluster other than a Cloudera cluster. Workaround: Before you run the mapping, create a table in the target database with columns corresponding to the connected ports.
OCON-7205 When you run a Sqoop mapping on the Blaze engine to export data of the Numeric data type from Netezza, the scale part of the data is truncated.
OCON-7078 Sqoop mappings that import data from or export data to an SSL-enabled database fail on the Blaze engine.
OCON-7076 When you run a Sqoop mapping and abort the mapping from the Developer tool, the Sqoop map- reduce jobs continue to run. Workaround: On the Sqoop data node, run the following command to kill the Sqoop map-reduce jobs: yarn application -kill
26 2020-06-22 Bug Description
OCON-688 When you enable Sqoop for a logical data object and export data to an IBM DB2 database, the Sqoop export command fails. However, the mapping runs successfully without any error. (456455)
IN-3362 A data object with an intelligent structure model might accept JSON input files formatted with newlines between attributes as suitable for partitioning, even though the files cannot be partitioned. This might cause data that corresponds to the model to be identified as lost data in the Spark log.
IDL-8652 If you run a mass ingestion specification that contains more than 2,000 tables, a majority of the ingestion mapping jobs fail with the following error: SEVERE: UserDefinedJob execution failed
BDM-9987 A mapping configured with a filter expression upstream of a Joiner transformation might be inconsistent with map-side joins.
BDM-6389 A mapping fails to add statistics to Hive table metadata after loading data to the table on Hortonworks. Workaround: To view statistics for a table, run the following command on the HIVE command line: ANALYZE TABLE