How Airbyte Works
About the source and destination
Apify Dataset
Apache Iceberg
Sync with Airbyte
Sync Manually
What sets Airbyte Apart
Modern GenAI Workflows
Move Large Volumes, Fast
An Extensible Open-Source Standard
Full Control & Security
Fully Featured & Integrated
Enterprise Support with SLAs
Apify is a web scraping and automation platform that can extract structured data from any website or automate any workflow on the web. For example, imagine you found a website selling shoes and want to get a spreadsheet with all the shoe sizes, colors, prices, etc., but the website doesn't make that information accessible in tabular form. Youcould certainly manually create such a spreadsheet using copy and paste, but that would take a lot of time and cause a lot of frustration. Or you can set up Apify to do this for you in a few seconds.
For huge analytical tables, Apache Iceberg is a high-performance format. Using Apache Iceberg, engines such as Spark, Trino, Flink, Presto, Hive and Impala can safely work with the same tables, at the same time, providing the reliability and simplicity of SQL tables to big data. With Apache Iceberg, you can merge new data, update existing rows, and delete specific rows. Data files can be eagerly rewritten or deleted deltas can be used to make updates faster.
1. First, navigate to the Apify website and log in to your account.
2. Once you are logged in, click on the "API" tab in the top navigation bar.
3. Next, click on the "Credentials" tab and then click the "Create new token" button.
4. Give your token a name and select the appropriate permissions for your use case.
5. Copy the generated token to your clipboard.
6. Navigate to your Airbyte dashboard and click on the "Sources" tab.
7. Click on the "Add Source" button and select "Apify" from the list of available connectors.
8. In the "Connection Configuration" section, paste the token you copied from Apify into the "API Token" field.
9. Enter the name of the dataset you want to connect to in the "Dataset Name" field.
10. Click the "Test" button to ensure that the connection is successful.
11. If the test is successful, click the "Save" button to save your configuration.
12. You can now use the Apify source connector in Airbyte to extract data from your chosen dataset.
1. Open the Airbyte platform and navigate to the "Destinations" tab on the left-hand side of the screen.
2. Click on the "Apache Iceberg" destination connector and select "Create new connection."
3. Enter a name for your connection and provide the necessary credentials for your Apache Iceberg database, including the host, port, database name, username, and password.
4. Test the connection to ensure that it is successful. 5. Select the tables or data sources that you want to replicate to your Apache Iceberg database.
6. Configure any additional settings or options for your connection, such as the frequency of data replication or any transformations that you want to apply to your data.
7. Save your connection and start the replication process.
8. Monitor the progress of your data replication and troubleshoot any issues that may arise.
9. Once the replication process is complete, verify that your data has been successfully replicated to your Apache Iceberg database.
10. Use your Apache Iceberg database to analyze and query your data as needed.
With Airbyte, creating data pipelines take minutes, and the data integration possibilities are endless. Airbyte supports the largest catalog of API tools, databases, and files, among other sources. Airbyte's connectors are open-source, so you can add any custom objects to the connector, or even build a new connector from scratch without any local dev environment or any data engineer within 10 minutes with the no-code connector builder.
We look forward to seeing you make use of it! We invite you to join the conversation on our community Slack Channel, or sign up for our newsletter. You should also check out other Airbyte tutorials, and Airbyte’s content hub!
What should you do next?
Hope you enjoyed the reading. Here are the 3 ways we can help you in your data journey:
What should you do next?
Hope you enjoyed the reading. Here are the 3 ways we can help you in your data journey:
Ready to get started?
Frequently Asked Questions
Apify's API provides access to a wide range of data types, including:
1. Web scraping data: Apify's web scraping tools allow users to extract data from websites and APIs, including HTML, JSON, XML, and CSV formats.
2. Social media data: Apify's API can be used to extract data from social media platforms such as Twitter, Facebook, and Instagram, including posts, comments, and user profiles.
3. E-commerce data: Apify's API can be used to extract data from e-commerce platforms such as Amazon, eBay, and Shopify, including product listings, prices, and reviews.
4. Search engine data: Apify's API can be used to extract data from search engines such as Google, Bing, and Yahoo, including search results, rankings, and keyword data.
5. Financial data: Apify's API can be used to extract financial data from sources such as stock exchanges, financial news websites, and investment platforms.
6. Weather data: Apify's API can be used to extract weather data from sources such as weather APIs and weather news websites.
7. Government data: Apify's API can be used to extract data from government websites and APIs, including census data, crime statistics, and public records.
Overall, Apify's API provides access to a wide range of data types, making it a powerful tool for data extraction and analysis.
What should you do next?
Hope you enjoyed the reading. Here are the 3 ways we can help you in your data journey: