Quick Start
Standard Cluster Deployment
Data Modeling
Connecting to The database
Data Migration
Data Query
Maintenance and Monitoring
Performance Tuning
Troubleshooting
Reference Guide
SQL Reference
FAQ
This document describes PXF-related frequently asked questions.
remote component error, Failed connect to localhost:5888; Connection refused (libchurl.c:950)
After PXF deployment, an error is reported when accessing HDFS:
remote component error,Failed connect to localhost:5888; Connection refused (libchurl.c:950)
Solution
The way PXF accesses files requires the PXF server to be enabled on the Master node, but the data file needs to be on the Segment PXF.
pxf/servers/core-site.xml and hdfs-site.xml
must be the same as the Hadoop configuration file.
pxf/servers/core-site.xml
configure user access rights.
The username and group of the file on Hadoop must be consistent with the specified pxf/core-site.xml
.
Solution
Add escape 'off'
to the option.
Use format 'text:multi'
.
OK.
Prerequisites
PXF can access S3 normally.
Solution
Copy the aa.csv
file and name it ab.csv
and upload it to bucket. The file path is ymatrix/test
. Currently, the ymatrix/test
files have aa.csv
and ab.csv
. The number of data strips in ymatrix/test
are 1,000.
DROP FOREIGN TABLE public.chen_test ;
CREATE FOREIGN TABLE public.chen_test (
c1 text,
c2 text,
c3 text
)
SERVER s3server_online
OPTIONS (
format 'csv',
resource 'ymatrix/test/a*.csv',
JSONIFY_ARRAY 'TRUE',
JSONIFY_MAP 'TRUE',
JSONIFY_RECORD 'TRUE'
);
SELECT count(*) FROM chen_test ;
count
-------
2000
(1 row)
Use the wildcard *
to recurse all files in all subdirectories under a directory.
Note! The file format that can be matched by wildcard characters must be the same, otherwise an error will be reported.
OK.
When files in TEXT
, CSV
, PARQUET
, and JSON
formats are automatically sliced according to the size of 128MB
slice and read in parallel by the corresponding segment according to the slice.