412 questions
0
votes
0
answers
26
views
Databricks external table lagging behind source files
I have a databricks external table which is pointed at an S3 bucket which contains an ever-growing number of parquet files (currently around 2000 of them). Each row in the file is timestamped to ...
0
votes
0
answers
73
views
Get row number from Databricks external table
If I have a external table set up in Databricks that points to a CSV file (or a folder full of CSV files), each record corresponds to a row in (one of) the file(s). Is there any way to determine the ...
-1
votes
1
answer
88
views
How to DELETE rows from source tables mapped to external tables in ELASTIC QUERY on Azure
Configuration
Database Engine = mybusiness.database.windows.net
Records Database = Records --> dbo.accesslogs
Legal Database = Legal --> dbo.personalinfo
Access Database = Access --> dbo....
0
votes
1
answer
70
views
Starburst unmanaged tables
I'm building a database on Starburst using unmanaged (or external) tables that read CSV files stored on S3.
The file path structure will be as follows:
bucket_name/file_name/year/month/day
The CSV ...
0
votes
1
answer
44
views
Load data in external table in Hive
I am new to hive just completed theory now doing some assignment, I am trying to load the csv data which looks like this : csv_file I have my HQL query I have checked the path for csv file is correct ...
0
votes
1
answer
189
views
How to set external location while using DLT and hive metastore
I originally set the Storage location in my DLT as abfss://{container}@{storageaccount}.dfs.core.windows.net/...
But when running the DLT I got the following error:
So I decided to leave the above ...
0
votes
2
answers
117
views
Can I access a row's file information in a Databricks external table?
I have an external table defined within my Databricks database that's pointing to a particular folder within an S3 bucket. There are multiple csv files in the folder, the contents of which all appear ...
0
votes
0
answers
44
views
Insert overwrite multiple partitions in an external Hive table
I am trying to overwrite multiple partitions in a large table. Basically I have my main external S3 table sandbox, partitioned by part:
scala> q("select * from sandbox")
+---+-------------...
1
vote
2
answers
252
views
Difference between Managed Table with a custom location and External Table in Databricks
What is the difference between a Managed Table with a location other than DBFS root (I created a schema with Location and created a managed table in that schema)and an External Table in Databricks?
2
votes
1
answer
154
views
External Table not returning values
I am trying to create external table in snowflake, which should be refreshed whenever we have a new file in the source directory.
External table is not producing any result even though my stage shows ...
0
votes
1
answer
274
views
Access External table on Synapse to access Delta table on Datalake
I'm trying to create an external table on Synapse serverless pool to access a delta table on ADLS gen 2.
I first created an external source as shown here.
--Create a database scoped credential using ...
0
votes
2
answers
102
views
Error Updating BigQuery Table from External Table
I have a BigQuery Table with about 200,000,000 rows. I have an external table that holds rows that need to have a value updated by having the row to be updated's unique id, and thestring that needs to ...
0
votes
2
answers
371
views
Bigquery: Create a temporary external table
According to the GCP official documentation
You can query an external data source in BigQuery by using a permanent table or a temporary table
https://cloud.google.com/bigquery/docs/external-tables#...
1
vote
1
answer
225
views
How create external table in Azure SQL to access a table in local SQL Server on same subnet. Getting error on select
I have a number of Azure SQL databases that I would like to have access an external table in a local SQL Server database (not named instance). The local SQL Server table has a trigger that can send ...
1
vote
1
answer
108
views
Why does Spark not create a new file after inserting data into an external table?
I have a .csv file data.csv stored at location: dbfs:/raw/data/externalTables/emp_data_folder/emp_data.csv
Here is a sample of the data in the file:
Alice,25,50000,North
Bob,30,60000,South
Charlie,35,...
0
votes
1
answer
55
views
can we load multiple file formats in vertica external table?
I am trying to load CSV and Parquet format files into one external table in Vertica DB but not finding any options, in worst case planning to load it in two different tables and make a union but union ...
0
votes
2
answers
165
views
ADX external tables fails, wrong managed identity
I am trying to create a new external table (kind=delta) in Azure Data Explorer.
My code is standard:
.create-or-alter external table external_someexport
(col1:int, col2:int, mydate:datetime )
kind=...
0
votes
1
answer
461
views
Databricks: Delete Delta log file using databricks SQL command
I have an external Delta table in Databricks with data files stored in an S3 backend. I want to completely remove(drop) this external Delta table, including all data files and Delta log files from the ...
1
vote
0
answers
44
views
String data with " as exterior quote character and \" as interior quote character causing error in GCP BQ Load
I have CSV data that looks like
id, xml
1, "{\"field1\": 1234, \"field2\": \"some string with a comma,\"}"
2, "{\"field1\": 5678, \"field3\&...
0
votes
1
answer
789
views
Is there a way to automatically generate a script in Azure SQL containing the create instructions for every table in the database
I have an existing data base in Azure that I need to replicate in another Azure database as External tables. To this end I need to make a script that creates all these 200+ tables.
I’m looking for a ...
0
votes
0
answers
735
views
BigQuery External Table Bad Rows Issue
I have a BigQuery table that's configured to be an External Table looking at Cloud Storage. The Source URI is:
gs://[bucketname]/test_file*
I did not specify maxBadRecords in my table creation request,...
1
vote
0
answers
228
views
I cannot REFRESH my External table at Snowflake
Im trying to refresh an external table but it is not working.
I need to REFRESH the table manually since I don't have an explicit periodicity to make it automatically. Also read the documentation: ...
1
vote
1
answer
427
views
Unable to create partitioned table in Databricks on external data residing on ADLS
I am copying files from ftp to ADLS in bronze folder under the partitioned folder heirarchy like
bronze
|
|
+----- __run_date__=2023-10-23
|
|
...
0
votes
1
answer
238
views
Cannot execute stored procedure in Oracle19c
I am currently facing an issue(below is the error code) when trying to import data from a .csv file to OracleDB. Now, I have a stored procedure so that I could easily call it out. Now, every time I ...
0
votes
1
answer
698
views
BigQuery external table using JSON files
I have e-commerce product data in JSON format. This data is frequently updated. I am planning to load this data into BigQuery. Given that the JSON files are huge (a few hundred attributes) and there ...
0
votes
1
answer
204
views
FILLER keyword throwing error in external table in oracle
While creating external table with SQL loader to load data, I used "FILLER" keyword specified as in oracle docs as below.
A filler field, specified by BOUNDFILLER or FILLER is a data file
...
0
votes
0
answers
340
views
How to get DDL statements of tables of external datasets in GCP BigQuery?
I am trying to take a clone of external dataset's tables using a script.
Before that I need to take out DDL statements of all tables and dataset.
In INFORMATION_SCHEMA.TABLES, DDL column is NULL for ...
0
votes
1
answer
137
views
ADX external tables - connect an ADLSgen2 Table `?
Is it possible to create a Kusto external table that is connected to an Azure storage account ?
The idea is not to connect a csv or parquet file in a container, but an ADLSgen2 Table like this:
....
0
votes
1
answer
167
views
How to change date value into EXTERNAL TABLE from a csv file
I need to populate my EXTERNAL TABLE from a .csv file received as input. Within my .csv file I have 3 fields which are respectively:
ORDER_DATE
-START
-END
Which are with the following format:
...
0
votes
1
answer
341
views
How to deal with columns containing commas while using comma as field terminator
I am creating an external table in Synapse serverless pool using data stored in a CSV file. I am using file format as DELIMITEDTEXT with Field Terminator as ",". Now one of my columns has ...
0
votes
0
answers
407
views
ORACLE 11G External Table Error ORA-29913
I created a folder named "datosext2" on C:\
I created a .txt file named "estados" on C:\datosext2
On sqldeveloper, I created a directory said folder
GRANT CREATE ANY DIRECTORY TO ...
0
votes
0
answers
319
views
MS SQL external table running very slow
I have one external table that has only 2 columns and 3 rows in it and when i run query to fetch data using external table that take almost 15 mins to get the data. I am not sure why that much slow as ...
0
votes
1
answer
429
views
Execute DDL to create a native table in BigQuery from files stored on GCS
I would like to create a native table in BigQuery from a CSV file with DDL statements in a way that replicates the "create table from..." feature in BigQuery's UI as shown in the screenshot ...
2
votes
1
answer
862
views
Snowflake external table that sits over csv files, not handling nulls as expected
I have many external tables that sits over csv files in S3. I have string values in numeric columns that should be parsed as null, e.g. 'na'. when I query the stage with the file format, this works as ...
0
votes
1
answer
108
views
Cannot create external database table with the DATA DIRECTORY clause
I would like to create a database table in an external file outside of the standard data directory.
The following works:
create table eh_extern (c1 integer) engine=innodb data directory = '/tmp';
I ...
0
votes
1
answer
738
views
Error creating an external table in BQ off partitioned parquet files
I'm trying to create an external table in BigQuery off partitioned parquet data found in GCS. I'm doing this by running the following SQL statement:
CREATE OR REPLACE EXTERNAL TABLE `project-id.poc....
0
votes
1
answer
993
views
If I add a column to a google sheet thats connected to big query will the big query external table get the new column?
I have a bigquery table that is connected to a google sheet link. I am going to add an additional column to this google sheet and wanted to know if the bigquery table schema will change to match? If ...
1
vote
0
answers
144
views
Redshift Spectrum returns null WITHOUT error
I'm querying redshift spectrum and certain fields are showing up null without any explanation. I've checked
SVL_S3LOG
SVL_SPECTRUM_SCAN_ERROR
SYS_EXTERNAL_QUERY_ERROR
And they are all empty.
In the ...
0
votes
1
answer
480
views
Polybase external data source for SQL Server
I'm having issues creating an external table in SQL Server which points to another SQL Server.
These are the steps I've taken:
CREATE EXTERNAL DATA SOURCE Polybase_ServerTest
WITH (
LOCATION = '...
0
votes
1
answer
1k
views
External table in Azure synapse - Error as parquet file is removed
I'm writing a parquet file to an azure data lake storage system, via databricks.
I'm using the following to write the parquet file:
cultural_split.write.partitionBy("filename").mode('...
0
votes
1
answer
1k
views
BigQuery Create External table from multiples files all files share same 18 columns, except newer files have one extra column
I have a situation I would like to know if possible to combat. I currently have a BigQueryUpsertTableOperator that goes to a google cloud storage and creates one external table from all the csv files ...
0
votes
1
answer
608
views
Kusto-use Managed Identity for External Table connection
I am trying to create an External Table in Azure Data Explorer using a user-assigned managed identity. The managed identity has been assigned to the cluster and I am using the statement as defined in ...
-1
votes
1
answer
112
views
Last Field in a CSV row left empty giving error: CSV table references column position 19, but line starting at position:325 contains only 19 columns
I have a DAG airflow task where it goes into a specific bucket and creates an external table from all the files inside together. I am getting the error (CSV table references column position 19, but ...
1
vote
1
answer
741
views
Getting Error: Missing close quote character (") when reading table after using BigQueryUpsertTableOperator from GCS bucket
I have a task that reads all files from a specific bucket path. All these files follow the same schema/format, and only difference is data is from each day so the file name is blahblah-(date).csv. I ...
0
votes
2
answers
684
views
Synapse CETAS from parquet file with columns definition is failing
In a Synapse pipeline, I'm trying to use a CETAS script activity from a parquet file that I generate before (from an Azure SQL database).
The source database has some tables containing LOB columns (...
1
vote
1
answer
541
views
How do I create an external table in Firebird and relate it to a CSV file?
I created a CSV file using a comma to separate fields. There are two field values on each line separated by a comma. I created an external table using two fields and referencing the filespec for the ...
0
votes
1
answer
2k
views
Azure Synapse > Serverless SQL > UPSERT parquet files is possible?
I'm using an Azure Synapse Serverless SQL database.
I have a first copy activity that uses Azure SQL tables as sources and an Azure storage gen2, where I store .parquet files as sink.
From these ....
1
vote
1
answer
161
views
Partition external table by unixtime
Typically when partitioning an external table, I will use 1-day boundaries with the syntax
.create external table extTbl (
Timestamp: datetime,
Data: dynamic
)
kind = blob
partition by (...
1
vote
2
answers
1k
views
big query external tables
There are many files which going to be land on gcs bucket daily. We have to merge them all and create a new file copied to new location.
I can do this with reading all the files which starts from ...
1
vote
1
answer
2k
views
Query a Snowflake table in Databricks as SQL - as external table
We have our primary data in snowflake and we want to share it with some users over in databricks.
The ask is not to "copy the data into databricks" as it will add to the cost. Rather just ...