Data export to S3 to redshift query

I exported parquet files to S3 successfully using the bulk export API

but from there - trying to query the data from Redshift things got tough , very tough.

I couldnt find a method that will query theentire run entry succesfully.

What I tried:

  1. using the copy command and define the table manually in redshift like the docs suggests (broken link there to AWS redshift copy command, btw).
  2. defining a Glue Crawler and Database and query those using Redshift spectrum, the schema should not be manually defined this way - that didnt work with an unhelpful error msg (if u ever worked with Redshift u know what kind of error msgs imtalking about)

My question is have anyone got thru this process succesfuly and was able to query the entire enrty from Redshift?

I suspect a few things

  1. could be that the spark version Glue is using under the hood is too old for some data types exported by langsmith.

  2. I try to buld the schema with the hive partitions (created by design by the langsmith exported, these are the folder tree of project/year/month/day - this is imoportant since the user must be able to filter by project which is not part of the langsmith’s run entry schema