Skip to main content
Filter by
Sorted by
Tagged with
Best practices
0 votes
5 replies
78 views

I have been working as a Data Engineer and got this issue. I came across a use case where I have a view(lets name it as inputView) which is created by reading data from some source. Now somewhere ...
Parth Sarthi Roy's user avatar
0 votes
0 answers
83 views

I created a table as follows: CREATE TABLE IF NOT EXISTS raw_data.civ ( date timestamp, marketplace_id int, ... some more columns ) USING ICEBERG PARTITIONED BY ( marketplace_id, ...
shiva's user avatar
  • 2,781
0 votes
0 answers
66 views

I am observing different write behaviors when executing queries on EMR Notebook (correct behavior) vs when using spark-submit to submit a spark application to EMR Cluster (incorrect behavior). When I ...
shiva's user avatar
  • 2,781
0 votes
0 answers
35 views

I'm analyzing Spark event logs and have already retrieved the SparkListenerStageSubmitted and SparkListenerTaskEnd events to collect metrics such as spill, skew ratio, memory, and CPU usage. However, ...
Carol C's user avatar
0 votes
0 answers
53 views

currently I'm working in a specific version of Apache Spark (3.1.1) that cannot upgrade. Since that I can't use Apache Sedona and the version 1.3.1 is too slow. My problem is the following code that ...
matdlara's user avatar
  • 149
0 votes
1 answer
115 views

I am trying to read the _delta_log folder of a delta lake table via spark to export some custom metrics. I have configured how to get some metrics from history and description but I have problem ...
Melika Ghiasi's user avatar
3 votes
0 answers
72 views

I am trying to write a custom decoder function in Java targeting Spark 4.0: public class MyDataToCatalyst extends UnaryExpression implements NonSQLExpression, ExpectsInputTypes, Serializable { //.....
Carsten's user avatar
  • 1,288
1 vote
0 answers
131 views

Spark reading data from MongoDB(ver 7.0) and DocumentDB(ver 4.0) and loading into the spark DataFrameReader is failing when DataFrameReader.isEmpty() method is called . SparkSession and ...
Sandeep Reddy CONT's user avatar
0 votes
1 answer
44 views

I have the below code where the Id is a 36 character GUID. The code gets executed but when a matching record is found , instead of updating it inserts the entire records again. What could be the root ...
Sandeep T's user avatar
  • 441
0 votes
1 answer
105 views

Its a sort of CDC ( Change Data Capture ) scenario in which I am trying to compare new data (in tblNewData) with old data (in tblOldData), and logging the changes into a log table (tblExpectedDataLog) ...
Aza's user avatar
  • 27
3 votes
1 answer
61 views

I am trying to implement the Parallelized BFS algorithm using Pyspark. I am following the material in CS246, What exactly in my implementation is making this thing take so long? Pardon me I am just a ...
Frenzy Ripper's user avatar
1 vote
1 answer
86 views

I'm encountering the following error while trying to upload a RocksDB checkpoint in Databricks: java.lang.IllegalStateException: Found no SST files during uploading RocksDB checkpoint version 498 with ...
Susmit Sarkar's user avatar
0 votes
1 answer
60 views

I am trying to calculate the timestamp difference on cumulative rows based on ID and status columns Example dataframe: ID TIMESTAMP STATUS V1 2023-06-18 13:00:00 1 V1 2023-06-18 13:01:00 1 V1 2023-06-...
RMK's user avatar
  • 41
0 votes
0 answers
62 views

I have the following code to test. I created a table on worker 1. Then I tried to read the table on worker 2 and it got TABLE_OR_VIEW_NOT_FOUND. Worker 2 is in the some computer as Master. I ran the ...
Rick C. Ferreira's user avatar
3 votes
0 answers
337 views

I upgraded PySpark from 3.5.5 to 3.5.6, and now all unit tests with an overwrite operation are failing with this error: pyspark.errors.exceptions.captured.AnalysisException: Table does not support ...
Nicholas Fiorentini's user avatar
1 vote
2 answers
111 views

I have a table containing the fields: user_ip, datetime, year, month, day, hour, tag_id, country, device_type, brand. I need to check if given a IP was active for a continuous period of 4 or more ...
user16798185's user avatar
0 votes
0 answers
101 views

I can make PySpark "work" no problem, but know very little and am very confused by documentation on performance. I have some source data partitioned by date, read it directory by directory (...
mateoc15's user avatar
  • 680
1 vote
1 answer
104 views

Here is minimal example using default data in DataBricks (Spark 3.4): import org.apache.spark.sql.functions.col import org.apache.spark.sql.{Row, SparkSession} import org.apache.spark.sql.types._ sc....
Igor Railean's user avatar
1 vote
2 answers
148 views

I have a table, base_df, with many columns, one of which is an array column: Id FruitNames Col1 Col2 Col3 ... Col99 1 ["apple", "banana", "orange"] ... ... ... ... ... 2 [...
wkeithvan's user avatar
  • 2,215
0 votes
0 answers
84 views

I am trying to deploy a scala application which uses structures streaming on a standalone distributed Spark cluster using the spark-submit command and I get the following error: Exception in thread &...
Maria's user avatar
  • 1
1 vote
0 answers
63 views

I have the following two datasets in Spark SQL: person view: person = spark.createDataFrame([ (0, "Bill Chambers", 0, [100]), (1, "Matei Zaharia", 1, [500, 250, 100]), (2, "...
DumbCoder's user avatar
  • 515
1 vote
0 answers
131 views

I am new to Pyspark and having few clarifications on writing dataframe to oracle database table using JDBC. As part of the requirement I need to read the data from Oracle table and perform ...
Siva's user avatar
  • 11
0 votes
0 answers
56 views

I use Spark to read JSON files that appear in a folder everyday with path pattern Yyyy/mm/dd to convert them into Iceberg format. Both folders JSON and Iceberg are in a s3 bucket on different paths. ...
Alex's user avatar
  • 1,019
0 votes
0 answers
65 views

I am experiencing data skew issues in spark, specifically during joins and window functions. I have tried many of the spark performance tuning configurations recommended but none appear to be working. ...
ifightfortheuserz's user avatar
1 vote
0 answers
132 views

I'm trying to create an Iceberg table with a geometry column in this example: import org.apache.sedona.sql.utils.SedonaSQLRegistrator SedonaSQLRegistrator.registerAll(spark) val stmt = ""&...
Stefan Ziegler's user avatar

1
2 3 4 5
539