You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Load test (can be manual one time , ad-hoc, no need for special tools, even could be functional test). Record results in this ticket), Try to ingest 1,000,000 and 10 million rows, record time and memory consumption.
Update README.md to indicate ingestion support and example (short code snippet)
The text was updated successfully, but these errors were encountered:
Note that the above is not the worst case scenario. The worst case scenario would be something like ingesting one million objects with the same schema, but randomized keys, e.g. one object would get a random number of keys from the schema. Let's say if we have an object with five keys, only 20% of objects will have all five keys, the others will have between one and four. This will trigger batching of the ingestion queries based on the keysets. Ingestion rows with the same keyset are batched together. This increases the number of queries, so it should theoretically be slower.
Support
job_input.send_object_for_ingestion(method="oracle")
orsend_tabular_data_for_ingestion(method='oracle')
If user passes method=oracle then the data would be insert in pre-configured oracle instance
So what needs to be done:
Create IngestToOracle which implements IIngesterPlugin
50% of the implementation is already done actually - in this branch - https://github.com/vmware/versatile-data-kit/blob/feature/vdk-oracle/projects/vdk-plugins/vdk-oracle/src/vdk/plugin/oracle/ingest_to_oracle.py . So use it as a starting point
Create functional tests that have an ingestion job . Make sure to cover all possible data types (that is have columns in the dataset with all types include datetime and decimal!!!) . Example job could be https://github.com/vmware/versatile-data-kit/blob/main/projects/vdk-plugins/vdk-trino/tests/test_ingest_to_trino.py . The tests should cover both cases into non-existing table (so auto-create works) and existing table (so it's updated)
Load test (can be manual one time , ad-hoc, no need for special tools, even could be functional test). Record results in this ticket), Try to ingest 1,000,000 and 10 million rows, record time and memory consumption.
Update README.md to indicate ingestion support and example (short code snippet)
The text was updated successfully, but these errors were encountered: