

Building your pipeline or Using Airbyte
Airbyte is the only open source solution empowering data teams to meet all their growing custom business demands in the new AI era.
- Inconsistent and inaccurate data
- Laborious and expensive
- Brittle and inflexible
- Reliable and accurate
- Extensible and scalable for all your needs
- Deployed and governed your way
Start syncing with Airbyte in 3 easy steps within 10 minutes



Take a virtual tour
Demo video of Airbyte Cloud
Demo video of AI Connector Builder
What sets Airbyte Apart
Modern GenAI Workflows
Move Large Volumes, Fast
An Extensible Open-Source Standard
Full Control & Security
Fully Featured & Integrated
Enterprise Support with SLAs
What our users say


"For TUI Musement, Airbyte cut development time in half and enabled dynamic customer experiences."


“Airbyte helped us accelerate our progress by years, compared to our competitors. We don’t need to worry about connectors and focus on creating value for our users instead of building infrastructure. That’s priceless. The time and energy saved allows us to disrupt and grow faster.”

"With Airbyte, we could just push a few buttons, allow API access, and bring all the data into Google BigQuery. By blending all the different marketing data sources, we can gain valuable insights."
Begin by ensuring you have access to both MS SQL Server and ClickHouse. Install and configure necessary client tools such as SQL Server Management Studio (SSMS) for MS SQL Server and clickhouse-client for ClickHouse. This will allow you to interact with both databases directly.
Use SQL Server Management Studio or command line tools to execute a query that extracts the data you need from MS SQL Server. For large datasets, consider exporting in chunks. You can export the data in CSV format using the SQL Server Import and Export Wizard or by running a query with the `bcp` (Bulk Copy Program) command line tool, like so:
```
bcp "SELECT * FROM YourTable" queryout "C:\path\to\output.csv" -c -t, -T -S servername\instance
```
Adjust the query and file path as necessary for your specific use case.
Ensure that the CSV file format aligns with ClickHouse's data import requirements. This might involve checking for proper delimiters (commas by default), handling of special characters, and ensuring consistent data types across rows. Open the CSV file in a text editor to verify its structure.
In ClickHouse, create a table that matches the schema of your MS SQL Server data. Use the ClickHouse client or a SQL query to define the table structure. For example:
```sql
CREATE TABLE your_table_name (
column1 DataType1,
column2 DataType2,
...
) ENGINE = MergeTree()
ORDER BY (column1);
```
Match the data types as closely as possible to ensure data integrity.
Move the CSV file to a location accessible by the ClickHouse server. This could be done using secure copy (SCP), FTP, or any file transfer mechanism that suits your environment. Ensure the ClickHouse server has read permissions for the file location.
Use the clickhouse-client to import the CSV data into ClickHouse. Run the following command, replacing placeholders with your specific details:
```bash
clickhouse-client --query="INSERT INTO your_table_name FORMAT CSV" < /path/to/output.csv
```
This command reads the CSV file and inserts the data into the specified ClickHouse table.
Finally, verify that the data has been correctly transferred by executing a few SELECT queries on the ClickHouse table. Check for data consistency, verify row counts, and ensure there are no discrepancies between the source and destination. This step is crucial to ensure the integrity and success of the data migration process.
By following these steps, you can manually move data from MS SQL Server to ClickHouse without relying on third-party connectors or integrations, ensuring a direct and controlled data transfer process.
FAQs
What is ETL?
ETL, an acronym for Extract, Transform, Load, is a vital data integration process. It involves extracting data from diverse sources, transforming it into a usable format, and loading it into a database, data warehouse or data lake. This process enables meaningful data analysis, enhancing business intelligence.
Microsoft SQL Server Consultants help companies choose the best business software solutions for their needs. Microsoft SQL Server Consultants help businesses resolve questions and issues, provide businesses with reliable information resources, and, ultimately, make better decisions on the software most appropriate for their unique needs. Consultants are available to help on call and can connect remotely to businesses’ computers to upgrade outdated editions of SQL servers to bring functions up to date for improved productivity.
MSSQL - SQL Server provides access to a wide range of data types, including:
1. Relational data: This includes tables, views, and stored procedures that are used to store and manipulate data in a structured format.
2. Non-relational data: This includes data that is not stored in a structured format, such as XML documents, JSON objects, and binary data.
3. Spatial data: This includes data that is related to geographic locations, such as maps, coordinates, and spatial queries.
4. Time-series data: This includes data that is related to time, such as timestamps, dates, and time intervals.
5. Graph data: This includes data that is related to relationships between entities, such as social networks, supply chains, and organizational structures.
6. Machine learning data: This includes data that is used for training and testing machine learning models, such as feature vectors, labels, and performance metrics.
7. Streaming data: This includes data that is generated in real-time, such as sensor data, log files, and social media feeds.
What is ELT?
ELT, standing for Extract, Load, Transform, is a modern take on the traditional ETL data integration process. In ELT, data is first extracted from various sources, loaded directly into a data warehouse, and then transformed. This approach enhances data processing speed, analytical flexibility and autonomy.
Difference between ETL and ELT?
ETL and ELT are critical data integration strategies with key differences. ETL (Extract, Transform, Load) transforms data before loading, ideal for structured data. In contrast, ELT (Extract, Load, Transform) loads data before transformation, perfect for processing large, diverse data sets in modern data warehouses. ELT is becoming the new standard as it offers a lot more flexibility and autonomy to data analysts.
What should you do next?
Hope you enjoyed the reading. Here are the 3 ways we can help you in your data journey: