Data Lake Insight
- Service Overview
- Getting Started
- DLI Console Overview
- SQL Editor
- Job Management
- Queue Management
- Data Management
- Job Templates
- Enhanced Datasource Connections
- Datasource Authentication
- Global Configuration
- Permissions Management
- Other Common Operations
- FAQ
- Flink Jobs
- Problems Related to SQL Jobs
- Problems Related to Spark Jobs
- Spark Jobs
- How Do I Use Spark to Write Data into a DLI Table?
- How Do I Set the AK/SK for a Queue to Operate an OBS Table?
- How Do I View the Resource Usage of DLI Spark Jobs?
- How Do I Use Python Scripts to Access the MySQL Database If the pymysql Module Is Missing from the Spark Job Results Stored in MySQL?
- How Do I Run a Complex PySpark Program in DLI?
- How Does a Spark Job Access a MySQL Database?
- How Do I Use JDBC to Set the spark.sql.shuffle.partitions Parameter to Improve the Task Concurrency?
- How Do I Read Uploaded Files for a Spark Jar Job?
- Why Are Errors "ResponseCode: 403" and "ResponseStatus: Forbidden" Reported When a Spark Job Accesses OBS Data?
- Why Is Error "verifyBucketExists on XXXX: status [403]" Reported When I Use a Spark Job to Access an OBS Bucket That I Have Access Permission?
- Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?
- Why Does the Job Fail to Be Executed and the Log Shows that the File Directory Is Abnormal When I Use a Spark Job to Access Files in SFTP?
- Why Does the Job Fail to Be Executed Due to Insufficient Database and Table Permissions?
- Why Can't I Find the Specified Python Environment After Adding the Python Package?
- Why Is a Spark Jar Job Stuck in the Submitting State?
- Product Consultation
- Quota
- Permission
- Queue
- Datasource Connections
- APIs
- SDKs
- Change History
- Help Center
- Data Lake Insight
- User Guide
- Service Overview
- Help Center
- Data Lake Insight
- Service Overview
- What Is Data Lake Insight?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- What Data Formats and Data Sources Are Supported by DLI Flink Jobs?
- Help Center
- Data Lake Insight
- FAQ
- Quota
- How Do I View My Quotas?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- How Do I Manage Fine-Grained DLI Permissions?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- Does the Description of a DLI Queue Can Be Modified?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Why Do I Need to Create a VPC Peering Connection for an Enhanced Datasource Connection?
- Help Center
- Data Lake Insight
- FAQ
- APIs
- Why Is Error "unsupported media Type" Reported When I Subimt a SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- SDKs
- How Do I Set the Timeout Duration for Querying SQL Job Results Using SDK?
- Help Center
- Data Lake Insight
- Job Management
- Spark Job Management
- Spark Job Management
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Help Center
- Data Lake Insight
- Job Templates
- Managing SQL Templates
- Help Center
- Data Lake Insight
- Job Templates
- Appendix
- TPC-H Sample Data in the SQL Template
- Help Center
- Data Lake Insight
- Global Configuration
- Global Variables
- Help Center
- Data Lake Insight
- Getting Started
- Creating and Submitting a Spark SQL Job
- Help Center
- Data Lake Insight
- Other Common Operations
- Importing Data to a DLI Table
- Help Center
- Data Lake Insight
- Getting Started
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Authorize a Subuser to View Flink Jobs?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Merge Small Files?
- Help Center
- Data Lake Insight
- Job Management
- SQL Job Management
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Use Spark to Write Data into a DLI Table?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Managing Flink Job Permissions
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Which Data Formats Does DLI Support?
- Help Center
- Data Lake Insight
- FAQ
- Quota
- How Do I Increase a Quota?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- What Is Column Permission Granting of a DLI Partition Table?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- Will Table Data in My Database Be Lost If I Delete a Queue?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Failed to Bind a Queue to an Enhanced Datasource Connection
- Help Center
- Data Lake Insight
- FAQ
- APIs
- Are Project IDs of Different Accounts the Same When They Are Used to Call APIs?
- Help Center
- Data Lake Insight
- FAQ
- SDKs
- How Do I Handle the dli.xxx,unable to resolve host address Error?
- Help Center
- Data Lake Insight
- Job Management
- Spark Job Management
- Creating a Spark Job
- Help Center
- Data Lake Insight
- Queue Management
- Queue Permission Management
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Managing Database Permissions
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Managing Permissions on Packages and Package Groups
- Help Center
- Data Lake Insight
- Job Templates
- Managing Flink Templates
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Cross-Source Analysis Development Methods
- Help Center
- Data Lake Insight
- Datasource Authentication
- Creating a CSS Datasource Authentication
- Help Center
- Data Lake Insight
- Global Configuration
- Permission Management for Global Variables
- Help Center
- Data Lake Insight
- Getting Started
- Developing and Submitting a Spark SQL Job Using the TPC-H Sample Template
- Help Center
- Data Lake Insight
- Permissions Management
- Creating an IAM User and Granting Permissions
- Help Center
- Data Lake Insight
- Other Common Operations
- Viewing Monitoring Metrics
- Help Center
- Data Lake Insight
- DLI Console Overview
- Help Center
- Data Lake Insight
- Service Overview
- Application Scenarios
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Set Auto Restart upon Exception for a Flink Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Specify an OBS Path When Creating an OBS Table?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Set the AK/SK for a Queue to Operate an OBS Table?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Preparing Flink Job Data
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- What Are the Differences Between MRS Spark and DLI Spark?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Why Does My Account Have Insufficient Permissions Due to Arrears?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Does DLI Ensure the Reliability of Spark Jobs When a Queue Is Abnormal?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- DLI Failed to Connect to GaussDB(DWS) Through an Enhanced Datasource Connection
- Help Center
- Data Lake Insight
- FAQ
- APIs
- What Can I Do If an Error Is Reported When the Execution of the API for Creating a SQL Job Times Out?
- Help Center
- Data Lake Insight
- Queue Management
- Creating a Queue
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Managing Table Permissions
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Creating a Package
- Help Center
- Data Lake Insight
- Job Templates
- Managing Spark SQL Templates
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Creating an Enhanced Datasource Connection
- Help Center
- Data Lake Insight
- Datasource Authentication
- Creating a Kerberos Datasource Authentication
- Help Center
- Data Lake Insight
- Global Configuration
- Service Authorization
- Help Center
- Data Lake Insight
- Getting Started
- Creating and Submitting a Spark Jar Job
- Help Center
- Data Lake Insight
- Permissions Management
- Creating a Custom Policy
- Help Center
- Data Lake Insight
- Other Common Operations
- DLI Operations That Can Be Recorded by CTS
- Help Center
- Data Lake Insight
- SQL Editor
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Save Flink Job Logs?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Create a Table Using JSON Data in an OBS Bucket?
- Help Center
- Data Lake Insight
- Job Management
- Spark Job Management
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I View the Resource Usage of DLI Spark Jobs?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Creating a Flink SQL Job
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Where Can DLI Data Be Stored?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Why Does the System Display a Message Indicating Insufficient Permissions When I Update a Program Package?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Do I Monitor Queue Exceptions?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Do I Do if the Datasource Connection Is Created But the Network Connectivity Test Fails?
- Help Center
- Data Lake Insight
- Queue Management
- Deleting a Queue
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Creating a Database or a Table
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Deleting a Package
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Deleting an Enhanced Datasource Connection
- Help Center
- Data Lake Insight
- Datasource Authentication
- Creating a Kafka_SSL Datasource Authentication
- Help Center
- Data Lake Insight
- Getting Started
- Creating and Submitting a Flink SQL Job
- Help Center
- Data Lake Insight
- Job Management
- Help Center
- Data Lake Insight
- Service Overview
- Permissions Management
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Can I Check Flink Job Results?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Set Local Variables in SQL Statements?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Use Python Scripts to Access the MySQL Database If the pymysql Module Is Missing from the Spark Job Results Stored in MySQL?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Creating a Flink Jar Job
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- What Are the Differences Between DLI Tables and OBS Tables?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Why Is Error "DLI.0003: Permission denied for resource..." Reported When I Run a SQL Statement?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Do I View DLI Queue Load?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Do I Configure the Network Between a DLI Queue and a Data Source?
- Help Center
- Data Lake Insight
- Queue Management
- Modifying the CIDR Block
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Deleting a Database or a Table
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Modifying the Owner
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Modifying Host Information
- Help Center
- Data Lake Insight
- Datasource Authentication
- Creating a Password Datasource Authentication
- Help Center
- Data Lake Insight
- Permissions Management
- DLI Request Conditions
- Help Center
- Data Lake Insight
- Queue Management
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error "No such user. userName:xxxx." Reported on the Flink Job Management Page When I Grant Permission to a User?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Can I Use the count Function to Perform Aggregation?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Run a Complex PySpark Program in DLI?
How Do I Run a Complex PySpark Program in DLI?
DLI natively supports PySpark.
For most cases, Python is preferred for data analysis, and PySpark is the best choice for big data analysis. Generally, JVM programs are packed into JAR files and depend on third-party JAR files. Similarly, Python programs also depend on third-party libraries, especially big data analysis programs related to PySpark-based converged machine learning. Traditionally, the Python library is installed on the execution machine based on pip. For serverless services such as DLI, you do not need to and are unaware of the underlying compute resources. In this case, how does DLI ensure that you run their programs perfectly?
DLI has built-in algorithm libraries for machine learning in its compute resources. These common algorithm libraries meet the requirements of most users. What if a user's PySpark program depends on a program library that is not provided by the built-in algorithm library? Actually, the dependency of PySpark is specified based on PyFiles. On the DLI Spark job page, you can directly select the Python third-party program library (such as ZIP and EGG) stored on OBS.
The compressed package of the dependent third-party Python library has structure requirements. For example, if the PySpark program depends on moduleA (import moduleA), the compressed package must meet the following structure requirement:

That is, the compressed package contains a folder named after a module name, and then the Python file of the corresponding class. Generally, the downloaded Python library may not meet this requirement. Therefore, you need to compress the Python library again. In addition, there is no requirement on the name of the compressed package. Therefore, it is recommended that you compress the packages of multiple modules into a compressed package. Now, a large and complex PySpark program is configured and runs normally.
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Debugging a Flink Job
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- How Can I Use DLI If Data Is Not Uploaded to OBS?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Why Can't I Query Table Data After I've Been Granted Table Permissions?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Do I Determine Whether There Are Too Many Jobs in the Current Queue?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- What Can I Do If a Datasource Connection Is Stuck in Creating State When I Try to Bind a Queue to It?
- Help Center
- Data Lake Insight
- Queue Management
- Elastic Queue Scaling
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Modifying the Owners of Databases and Tables
- Help Center
- Data Lake Insight
- Data Management
- Package Management
- Built-in Dependencies
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Binding an Elastic Resource Pool
- Help Center
- Data Lake Insight
- Datasource Authentication
- Datasource Authentication Permission Management
- Help Center
- Data Lake Insight
- Permissions Management
- Common Operations Supported by DLI System Policy
- Help Center
- Data Lake Insight
- Data Management
- Help Center
- Data Lake Insight
- Service Overview
- Basic Concepts
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Know Which Checkpoint the Flink Job I Stopped Will Be Restored to When I Start the Job Again?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Synchronize DLI Table Data from One Region to Another?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Does a Spark Job Access a MySQL Database?
- Help Center
- Data Lake Insight
- Job Management
- Flink Job Management
- Performing Operations on a Flink Job
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Can I Import OBS Bucket Data Shared by Other Tenants into DLI?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Will an Error Be Reported if the Inherited Permissions Are Regranted to a Table That Inherits Database Permissions?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Do I Switch an Earlier-Version Spark Queue to a General-Purpose Queue?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Do I Connect DLI to Data Sources?
- Help Center
- Data Lake Insight
- Queue Management
- Scheduling CU Changes
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Importing Data to the Table
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Unbinding an Elastic Resource Pool
- Help Center
- Data Lake Insight
- Job Templates
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is a Message Displayed Indicating That the SMN Topic Does Not Exist When I Use the SMN Topic in DLI?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Insert Table Data into Specific Fields of a Table Using a SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Use JDBC to Set the spark.sql.shuffle.partitions Parameter to Improve the Task Concurrency?
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Why Is Error "Failed to create the database. {"error_code":"DLI.1028";"error_msg":"Already reached the maximum quota of databases:XXX"." Reported?
- Help Center
- Data Lake Insight
- FAQ
- Permission
- Why Can't I Query a View After I'm Granted the Select Table Permission on the View?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- Why Cannot I View the Resource Running Status of DLI Queues on Cloud Eye?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Can I Perform Query on Data Stored on Services Rather Than DLI?
- Help Center
- Data Lake Insight
- Queue Management
- Testing Address Connectivity
- Help Center
- Data Lake Insight
- Data Management
- Databases and Tables
- Exporting Data from DLI to OBS
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Much Data Can Be Processed in a Day by a Flink SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "path obs://xxx already exists" Reported When Data Is Exported to OBS?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- How Do I Read Uploaded Files for a Spark Jar Job?
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Can a Member Account Use Global Variables Created by Other Member Accounts?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- How Do I Allocate Queue Resources for Running Spark Jobs If I Have Purchased 64 CUs?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Can I Access Data Across Regions?
- Help Center
- Data Lake Insight
- Queue Management
- Creating an SMN Topic
- Help Center
- Data Lake Insight
- Datasource Authentication
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Does Data in the Temporary Stream of Flink SQL Need to Be Cleared Periodically? How Do I Clear the Data?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "SQL_ANALYSIS_ERROR: Reference 't.id' is ambiguous, could be: t.id, t.id.;" Displayed When Two Tables Are Joined?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Are Errors "ResponseCode: 403" and "ResponseStatus: Forbidden" Reported When a Spark Job Accesses OBS Data?
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- How Do I Manage Tens of Thousands of Jobs Running on DLI?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- Why Is Error "Queue plans create failed. The plan xxx target cu is out of quota" Reported When I Schedule CU Changes?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- How Do I Set the Auto-increment Primary Key or Other Fields That Are Automatically Filled in the RDS Table When Creating a DLI and Associating It with the RDS Table?
- Help Center
- Data Lake Insight
- Queue Management
- Managing Queue Tags
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Enhanced Connection Permission Management
- Help Center
- Data Lake Insight
- Global Configuration
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is a Message Displayed Indicating That the OBS Bucket Is Not Authorized When I Select an OBS Bucket for a Flink SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "The current account does not have permission to perform this operation,the current account was restricted. Restricted for no budget." Reported when a SQL Statement Is Executed?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Is Error "verifyBucketExists on XXXX: status [403]" Reported When I Use a Spark Job to Access an OBS Bucket That I Have Access Permission?
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- How Do I Change the Name of a Field in a Created Table?
- Help Center
- Data Lake Insight
- FAQ
- Queue
- Why Is a Timeout Exception Reported When a DLI SQL Statement Fails to Be Executed on the Default Queue?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Why Is the Error Message "communication link failure" Displayed When I Use a Newly Activated Datasource Connection?
- Help Center
- Data Lake Insight
- Enhanced Datasource Connections
- Enhanced Datasource Connection Tag Management
- Help Center
- Data Lake Insight
- Permissions Management
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Create an OBS Partitioned Table for a Flink SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "There should be at least one partition pruning predicate on partitioned table XX.YYY" Reported When a Query Statement Is Executed?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?
- Help Center
- Data Lake Insight
- FAQ
- Product Consultation
- Does DLI Have the Apache Spark Command Injection Vulnerability (CVE-2022-33891)?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Connection Times Out During MRS HBase Datasource Connection, and No Error Is Recorded in Logs
- Help Center
- Data Lake Insight
- Other Common Operations
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Dump Data to OBS and Create an OBS Partitioned Table?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "IllegalArgumentException: Buffer size too small. size" Reported When Data Is Loaded to an OBS Foreign Table?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Does the Job Fail to Be Executed and the Log Shows that the File Directory Is Abnormal When I Use a Spark Job to Access Files in SFTP?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Why Can't I Find the Subnet When Creating a DLI Datasource Connection?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error Message "DLI.0005" Displayed When I Use an EL Expression to Create a Table in a Flink SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "DLI.0002 FileNotFoundException" Reported During SQL Job Running?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Does the Job Fail to Be Executed Due to Insufficient Database and Table Permissions?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Error Message "Incorrect string value" Is Displayed When insert overwrite Is Executed on a Datasource RDS Table
- Help Center
- Data Lake Insight
- Change History
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is No Data Queried in the DLI Table Created Using the OBS File Path When Data Is Written to OBS by a Flink Job Output Stream?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is a Schema Parsing Error Reported When I Create a Hive Table Using CTAS?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Can't I Find the Specified Python Environment After Adding the Python Package?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Null Pointer Error Is Displayed When the System Creates a Datasource RDS Table
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Does a Flink SQL Job Fails to Be Executed, and Is "connect to DIS failed java.lang.IllegalArgumentException: Access key cannot be null" Displayed in the Log?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "org.apache.hadoop.fs.obs.OBSIOException" Reported When I Run DLI SQL Scripts on DataArts Studio?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to Spark Jobs
- Why Is a Spark Jar Job Stuck in the Submitting State?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- Error Message "org.postgresql.util.PSQLException: ERROR: tuple concurrently updated" Is Displayed When the System Executes insert overwrite on a Datasource GaussDB(DWS) Table
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error "Not authorized" Reported When a Flink SQL Job Reads DIS Data?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "UQUERY_CONNECTOR_0001:Invoke DLI service api failed" Reported in the Job Log When I Use CDM to Migrate Data to DLI?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- RegionTooBusyException Is Reported When Data Is Imported to a CloudTable HBase Table Through a Datasource Table
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Data Writing Fails After a Flink SQL Job Consumed Kafka and Sank Data to the Elasticsearch Cluster
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "File not Found" Reported When I Access a SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- A Null Value Is Written Into a Non-Null Field When a DLI Datasource Connection Is Used to Connect to a GaussDB(DWS) Table
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Configure Checkpoints for Flink Jar Jobs and Save the Checkpoints to OBS?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "DLI.0003: AccessControlException XXX" Reported When I Access a SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Datasource Connections
- An Insert Operation Failed After the Schema of the GaussDB(DWS) Source Table Is Updated
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Does a Flink JAR Job Support Configuration File Upload? How Do I Upload a Configuration File?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "DLI.0001: org.apache.hadoop.security.AccessControlException: verifyBucketExists on {{bucket name}}: status [403]" Reported When I Access a SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Does the Submission Fail Due to Flink JAR File Conflict?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is Error "The current account does not have permission to perform this operation,the current account was restricted. Restricted for no budget" Reported During SQL Statement Execution? Restricted for no budget.
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Does a Flink Jar Job Fail to Access GaussDB(DWS) and a Message Is Displayed Indicating Too Many Client Connections?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Troubleshoot Slow SQL Jobs?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error Message "Authentication failed" Displayed During Flink Jar Job Running?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I View DLI SQL Logs?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error Invalid OBS Bucket Name Reported After a Flink Job Submission Failed?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I View SQL Execution Records?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Does the Flink Submission Fail Due to Hadoop JAR File Conflict?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Eliminate Data Skew by Configuring AE Parameters?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Connect a Flink jar Job to SASL_SSL?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- What Can I Do If a Table Cannot Be Queried on the DLI Console?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Optimize Performance of a Flink Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- The Compression Ratio of OBS Tables Is Too High
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Write Data to Different Elasticsearch Clusters in a Flink Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Can I Avoid Garbled Characters Caused by Inconsistent Character Codes?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Prevent Data Loss After Flink Job Restart?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Do I Need to Grant Table Permissions to a User and Project After I Delete a Table and Create One with the Same Name?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Locate a Flink Job Submission Error?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Can't I Query Table Data After Data Is Imported to a DLI Partitioned Table Because the File to Be Imported Does Not Contain Data in the Partitioning Column?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Locate a Flink Job Running Error?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- How Do I Fix the Data Error Caused by CRLF Characters in a Field of the OBS File Used to Create an External OBS Table?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- How Do I Know Whether a Flink Job Can Be Restored from a Checkpoint After Being Restarted?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Does a SQL Job That Has Join Operations Stay in the Running State?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Does DIS Stream Not Exist During Job Semantic Check?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- The on Clause Is Not Added When Tables Are Joined. Cartesian Product Query Causes High Resource Usage of the Queue, and the Job Fails to Be Executed
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is the OBS Bucket Selected for Job Not Authorized?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Can't I Query Data After I Manually Add Data to the Partition Directory of an OBS Table?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Are Logs Not Written to the OBS Bucket After a DLI Flink Job Fails to Be Submitted for Running?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is All Data Overwritten When insert overwrite Is Used to Overwrite Partitioned Table?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Information Displayed on the FlinkUI/Spark UI Page Incomplete?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is a SQL Job Stuck in the Submitting State?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is the Flink Job Abnormal Due to Heartbeat Timeout Between JobManager and TaskManager?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is the create_date Field in the RDS Table Is a Timestamp in the DLI query result?
- Help Center
- Data Lake Insight
- FAQ
- Flink Jobs
- Why Is Error "Timeout expired while fetching topic metadata" Repeatedly Reported in Flink JobManager Logs?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- What Can I Do If datasize Cannot Be Changed After the Table Name Is Changed in a Finished SQL Job?
- Help Center
- Data Lake Insight
- FAQ
- Problems Related to SQL Jobs
- Why Is the Data Volume Changes When Data Is Imported from DLI to OBS?