MongoDB is a popular open-source NoSQL database that stores data in a flexible, document-based format. It is designed to handle large volumes of unstructured data and is highly scalable, making it a popular choice for modern web applications. MongoDB uses a JSON-like format to store data, which allows for easy integration with web applications and APIs. It also supports dynamic queries, indexing, and aggregation, making it a powerful tool for data analysis. MongoDB is widely used in industries such as finance, healthcare, and e-commerce, and is known for its ease of use and flexibility.
Typesense is an open-source, typo-tolerant search engine optimized for an instant (typically sub-50ms) search-like-up-type experience and developer productivity. If you've heard of Elasticsearch or Algolia, a good way to think about Typesense is that it's an open source alternative to Algolia, with some key issues fixed and an easy-to-use battery-powered alternative to Elasticsearch.It works like a CDN, but for Search. Deploy nodes around the world, closest to your users, to provide them an ultra-fast search experience.

1. First, you need to have a MongoDB instance running and accessible from the internet. You will also need to have the necessary credentials to access the database.
2. In the Airbyte dashboard, click on "Sources" and then click on "New Source."
3. Select "MongoDB" from the list of available sources.
4. In the "Connection Configuration" section, enter the following information:
- Host: The hostname or IP address of your MongoDB instance.
- Port: The port number on which your MongoDB instance is running.
- Username: The username you use to access your MongoDB instance.
- Password: The password you use to access your MongoDB instance.
- Authentication Database: The name of the database where your authentication credentials are stored.
5. Click on "Test Connection" to ensure that Airbyte can connect to your MongoDB instance.
6. If the connection is successful, click on "Save" to save your MongoDB source configuration.
7. You can now create a new pipeline and select your MongoDB source as the input. You can then configure the pipeline to transform and load your data into your desired destination.

1. First, navigate to the Airbyte website and log in to your account.
2. Once you are logged in, click on the "Destinations" tab on the left-hand side of the screen.
3. Scroll down until you find the Typesense destination connector and click on it.
4. You will be prompted to enter your Typesense API key. Enter this information and click "Test Connection" to ensure that the connection is successful.
5. If the connection is successful, click "Save" to save your Typesense destination connector settings.
6. Next, navigate to the "Sources" tab on the left-hand side of the screen and select the source that you want to connect to your Typesense destination.
7. Follow the prompts to enter the necessary information for your source connector, such as the API key or database credentials.
8. Once you have entered all of the necessary information, click "Test Connection" to ensure that the connection is successful.
9. If the connection is successful, click "Save" to save your source connector settings.
10. Finally, click on the "Sync" tab on the left-hand side of the screen and select the source and destination connectors that you want to use for your data sync.
11. Follow the prompts to set up your data sync, such as selecting the tables or data types that you want to sync.
12. Once you have completed all of the necessary steps, click "Start Sync" to begin syncing your data between your source and Typesense destination connectors.

With Airbyte, creating data pipelines take minutes, and the data integration possibilities are endless. Airbyte supports the largest catalog of API tools, databases, and files, among other sources. Airbyte's connectors are open-source, so you can add any custom objects to the connector, or even build a new connector from scratch without any local dev environment or any data engineer within 10 minutes with the no-code connector builder.
We look forward to seeing you make use of it! We invite you to join the conversation on our community Slack Channel, or sign up for our newsletter. You should also check out other Airbyte tutorials, and Airbyte’s content hub!
Ready to get started?
Frequently Asked Questions
What is ETL?
ETL, an acronym for Extract, Transform, Load, is a vital data integration process. It involves extracting data from diverse sources, transforming it into a usable format, and loading it into a database, data warehouse or data lake. This process enables meaningful data analysis, enhancing business intelligence.
MongoDB gives access to a wide range of data types, including:
1. Documents: MongoDB stores data in the form of documents, which are similar to JSON objects. Each document contains a set of key-value pairs that represent the data.
2. Collections: A collection is a group of related documents that are stored together in MongoDB. Collections can be thought of as tables in a relational database.
3. Indexes: MongoDB supports various types of indexes, including single-field, compound, and geospatial indexes. Indexes are used to improve query performance.
4. GridFS: MongoDB's GridFS is a specification for storing and retrieving large files, such as images and videos, in MongoDB.
5. Aggregation: MongoDB's aggregation framework provides a way to perform complex data analysis operations, such as grouping, filtering, and sorting, on large datasets.
6. Transactions: MongoDB supports multi-document transactions, which allow multiple operations to be performed atomically.
7. Change streams: MongoDB's change streams provide a way to monitor changes to data in real-time, allowing applications to react to changes as they occur.
Overall, MongoDB provides access to a flexible and powerful data model that can handle a wide range of data types and use cases.
What is ELT?
ELT, standing for Extract, Load, Transform, is a modern take on the traditional ETL data integration process. In ELT, data is first extracted from various sources, loaded directly into a data warehouse, and then transformed. This approach enhances data processing speed, analytical flexibility and autonomy.
Difference between ETL and ELT?
ETL and ELT are critical data integration strategies with key differences. ETL (Extract, Transform, Load) transforms data before loading, ideal for structured data. In contrast, ELT (Extract, Load, Transform) loads data before transformation, perfect for processing large, diverse data sets in modern data warehouses. ELT is becoming the new standard as it offers a lot more flexibility and autonomy to data analysts.