URL file dialog
The URL File Dialog is used to navigate through the file system and select input or output files.
In many components, you are asked to specify the URL of some files. These files can serve to locate the sources of data that should be read, the sources to which data should be written or the files that must be used to transform data flowing through a component and some other file URL. To specify the URL of such a file, you can use the URL File Dialog.
To access the URL File Dialog, double-click on a component, click the Filter expression row in the component editor and then the 3-point button on the right.
The URL File Dialog has several tabs on it.
Local Files
Use the Local files tab to locate files on a local file system. The combo contains local file system places and parameters. It can be used to specify both Data Shaper projects and any other local files.
Best practice is to specify the path to files with Workspace view instead of Local view. Workspace view with help of parameters provides you with better portability of your graphs.
Workspace View
Workspace view tab serves to locate files in a workspace of a local Data Shaper project.
Data Shaper Server
Data Shaper Server dialog serves to locate files of all opened Data Shaper Server projects. Available only for Data Shaper Server projects.
Hadoop HDFS
Use the Hadoop HDFS tab to locate files on Hadoop Distributed File System.
You need a working Hadoop Connection to choose the particular files.
Remote Files
The Remote files tab serves to locate files on a remote computer or on the Internet. You can specify properties of connection, proxy settings, and HTTP properties.
You can type the URL directly in the format described in Supported File URL Formats for Readers or Supported File URL Formats for Writers, or you can specify it with a help of Edit URL Dialog. The Edit URL Dialog is accessible under the icon .
Edit URL Dialog
Edit URL Dialog lets you specify connection to a remote server in an easy way. Choose the protocol, specify a host name, port, credentials, and path.
The dialog lets you specify the connection using the following protocols:
-
HTTP
-
HTTPS
-
FTP
-
SFTP - FTP over SSH
-
Amazon S3
-
Azure Blob Storage
-
WebDav
-
WebDav over SSL
-
Windows Share - SMB1/CIFS
-
Windows Share - SMB 2.x, SMB 3.x
Click Save to save the connection settings. Click OK to use it.
The Load button serves to load a session from the list for subsequent editing.
The Delete button serves to delete the session from the list.
HTTP(S), (S)FTP, WebDav, and SMB
If the protocol is HTTP, HTTPS, FTP, SFTP - FTP over SSH, WebDav, WebDav over SSL, Windows Share - SMB1/CIFS or Windows Share - SMB 2.x or 3.x, the dialog allows you to specify the host name, port, username, password, and path on the server. It allows you to connect anonymously, as well.
SFTP Certificate in Data Shaper
If you are reading from or writing into remote files and are connected via an SFTP protocol using a certificate-based authorization, you should do one of the following:
- Option 1: Create an OpenSSH configuration file and specify the path to it in the Preferences (in the Designer go to Window > Preferences) as per the screenshot below. The configuration file can hold multiple configurations for different hosts.
- Option 2: Create a directory named
ssh-keys
in your project, and put the private key files into this directory and choose a suitable filename with the.key
suffix. Listed in order from the highest to lowest priority when resolving, the private key file can have the following names:
a.[email protected]
b.hostname.key
c.*.key
(the files are resolved in alphabetical order).
Hint!
If you want to explicitly select a certificate for a specific location, the best way is to use the name with the highest priority, i.e. [email protected]. In such a case, if the connection succeeds, other keys are ignored.
Figure below shows the format of the OpenSSH private key generated by ssh-keygen
.
URL Syntax for FTP Proxy
Data Shaper is able to connect to FTP proxy using the following URL syntax:
ftp://username%40proxyuser%40ftphost:password%40proxypassword@proxyhost
where:
username
- Your login on the FTP server.
proxyuser
- Your login on the proxy server.
ftphost
- The hostname of the FTP server.
password
- Your FTP password.
proxypassword
- Your proxy password.
proxyhost
- The hostname of the proxy server.
Amazon S3
In the case of the Amazon S3 protocol, the dialog allows you to fill in access Key, secret key, bucket, and path. For better performance, you should fill in the corresponding region.
Having the connection specified, you can choose the particular file(s).
Amazon S3 URL
It is recommended to connect to S3 via endpoint-specific S3 URL: s3://s3.eu-central-1.amazonaws.com/bucket.name/
. The end-point in URL should be the end-point corresponding to the bucket.
-
The URL with a specific endpoint has a much better performance than the generic one (
s3://s3.amazonaws.com/bucket.name/
), but you can only access the buckets of the specific region. -
The endpoint affects the signature version that will be used. If you connect to the generic one, the signature version may not match the endpoint being used. Therefore the signature is sent twice and you can see an error message in the error log:
DEBUG [main] - Received error response: com.amazonaws.services.s3.model.AmazonS3Exception: The authorization mechanism you have provided is not supported. Please use AWS4-HMAC-SHA256. (Service: null; Status Code: 400; Error Code: InvalidRequest; Request ID: 2D7C4933BD5ED2F8), S3 Extended Request ID: 9wmejqgrZ0jRpgqvw43RXUBZOzm9rnd5/wVN19kSe0dHAF/k5rxq34jvRhy8bHd5JnqBcQTBwkM=
WARN [main] - Attempting to re-send the request to cloverdx.example.test.s3.eu-central-1.amazonaws.com with AWS V4 authentication. To avoid this warning in the future, please use region-specific endpoint to access buckets located in regions that require V4 signing.
For list of regions and endpoints, see AWS Regions and Endpoints (Amazon S3).
When the S3 URL does not contain Secret Key + Access Key (e.g. s3://s3.eu-central-1.amazonaws.com/bucket.name/path
), Data Shaper automatically searches for credentials in the following sources (in this order):
- Environment Variables
AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY
Recommended since they are recognized by all the AWS SDKs and CLI except for .NETAWS_ACCESS_KEY and AWS_SECRET_KEY
only recognized by Java SDK
-
Java System Properties -
aws.accessKeyId
andaws.secretKey
-
Credential profiles file at the default location (
~/.aws/credentials
)
shared by all AWS SDKs and the AWS CLI -
Credentials delivered through the Amazon EC2 container service
theAWS_CONTAINER_CREDENTIALS_RELATIVE_URI
environment variable must be set and the security manager must have permission to access the variable -
Instance profile credentials delivered through the Amazon EC2 metadata service
For detailed information, see the Walkthrough: Using IAM roles for EC2 instances.
Capt. Eddie to ground control:"These sources of credentials may be used for graph development in a local project; for example, set
aws.accessKeyId
andaws.secretKey
Java system properties (for Data Shaper Runtime) and add them toCloverDXDesigner.ini
(for File URL dialog) so that graphs work in local projects when using S3 URLs without credentials."
Azure Blob Storage
Microsoft Azure Blob Storage is a cloud object storage service, similar to Amazon S3. Data Shaper supports Azure Blob Storage since version 5.11.
There are multiple supported authentication schemes:
- Storage Shared Key
https://docs.microsoft.com/en-us/rest/api/storageservices/authorize-with-shared-key
This authentication is the easiest to set up. It is similar to username/password authentication. You use the name of the storage account as the username and the Access Key as the password. The disadvantage is that all applications that use the Access Key have the same permissions.
You can find the key here: Azure Portal - Storage accounts - - Access keys
az-blob://[account]:[key]@[account].blob.core.windows.net/container/path
oraz-blob://AccountName=[account]:AccountKey=[key]@[account].blob.core.windows.net/container/path
to avoid confusion with the Client Secret authentication.
Note that the key must be URL-encoded before you can use it in the URL. The Edit URL dialog encodes the key automatically.
Example
Plain key:XFqGQY9/FRBucrRKldxykYUp9WmnzFHR9to/w2sP9+fXoDAKoTfWvdUOAzcaS3Wnon9mIgRbPcudtlwsNPtwzQ==
Encoded key:XFqGQY9%2FFRBucrRKldxykYUp9WmnzFHR9to%2Fw2sP9%2BfXoDAKoTfWvdUOAzcaS3Wnon9mIgRbPcudtlwsNPtwzQ%3D%3D
- Client Secret
https://docs.microsoft.com/en-us/azure/container-registry/container-registry-authentication#service-principal
This authentication scheme allows fine-grained access control, because you can set different permissions for each application that uses your storage.
First, create an "application" for your Data Shaper processing in your Azure Active Directory: Azure Portal - Azure Active Directory - App registrations
The authentication scheme uses three values: Tenant ID, Client ID (also called Application ID) and Client Secret.
You can find the Tenant ID and Client ID in the Overview of your application.
The Client Secret is in the Certificates & secrets section of your application.
Create a new secret and copy the Value, not the Secret ID.
az-blob://TenantId=[TenantId]:ClientId=[ClientId]:ClientSecret=[ClientSecret]@[account].blob.core.windows.net
or justaz-blob://[TenantId]:[ClientId]:[ClientSecret]@[account].blob.core.windows.net
- Environment Variables
Instead of putting the authentication information into the URL, you can configure the connection using the environment variables below.
The URL then contains only the storage account as a part of the host name:
az-blob://[account].blob.core.windows.net/container/path
-
Connection String
https://docs.microsoft.com/en-us/azure/storage/common/storage-configure-connection-string
You can find the connection string next to your Access Key: Azure Portal - Storage accounts - - Access keys- AZURE_STORAGE_CONNECTION_STRING
Example
export AZURE_STORAGE_CONNECTION_STRING="DefaultEndpointsProtocol=https;AccountName=[account];AccountKey=XFqGQY9/FRBucrRKldxykYUp9WmnzFHR9to/w2sP9+fXoDAKoTfWvdUOAzcaS3Wnon9mIgRbPcudtlwsNPtwzQ==;EndpointSuffix=core.windows.net"
- AZURE_STORAGE_CONNECTION_STRING
-
Client Secret
See Client Secret Authentication above.- AZURE_CLIENT_ID
- AZURE_CLIENT_SECRET
- AZURE_TENANT_ID
-
Client Certificate
You can also set up certificates in the Certificates & secrets section of your application in Azure Active Directory.- AZURE_CLIENT_ID
- AZURE_TENANT_ID
- AZURE_CLIENT_CERTIFICATE_PATH
-
Username and Password
- AZURE_CLIENT_ID
- AZURE_USERNAME
- AZURE_PASSWORD
-
Managed Identity
If the application is deployed to an Azure host with Managed Identity enabled, Data Shaper will authenticate with that account.
az-blob://[account].blob.core.windows.net/container/path
-
Anonymous
If none of the above applies, Data Shaper attempts to connect anonymously.
Anonymous access must be explicitly enabled on the container. Clients can then read data from the container without authorization.
az-blob://[account].blob.core.windows.net/container/path
Port
Serves to specify fields and processing type for port reading or writing. Opens only in components that allow such data source or target.
See also: Input Port Reading or Output Port Writing
Dictionary
Dictionary tab serves to specify dictionary key value and processing type for dictionary reading or writing. Opens only in components that allow such data source or target.
See also: Using a Dictionary in Graphs
Filtering Files and Tips
If you use File URL Dialog configured to display only some files according to the extension, you can see the File Extension below File URL.
More detailed information of URLs for each of the tabs described above is provided in sections:
Updated 5 months ago