sql

Making SQL Queries with MarkLogic

Reading Time: 3 minutes MarkLogic is the only Enterprise NoSQL database, that brings all the features you need into one unified system. SQL is one of the many languages that has been supported by MarkLogic. So that data stored in a MarkLogic database can be easily consumed by technologies that require a SQL data source. SQL queries can be sent via server-side applications code or through ODBC. The view is required so Continue Reading

knoldus data-carbon

How To Create And Insert Data In Postgre Hstore

Reading Time: 2 minutes Postgre without a doubt is one of the most popular databases in the market. The reason behind this is PostgreSQL’s speed, security, and robustness. PostgreSQL isn’t just a relational database. PostgreSQL invented many post-relational concepts like in-database functions and user-defined datatypes. Other than that it has many amazing features like Hstore. Now, this blog is for you if you want to know about PostgreSQL Hstore. Continue Reading

BigQuery: Querying nested arrays

Reading Time: 2 minutes In a previous blog, we had seen BigQuery facilitate efficient data warehouse schema design. BigQuery supports the nested & repeated columns. We can use a combination of ARRAY and STRUCT data types to define our schema in BigQuery. It enables to denormalize data efficiently in single table. In this blog, for the same schema of sales data, we will execute a few DML operations on nested array fields. Schema In Continue Reading

Let’s have a look at the PostgreSQL CRUD operation

Reading Time: 4 minutes We are aware that PostgreSQL is one of the most popular and powerful databases in the market. Its amazing features and support from large developer communities always make a difference in the industry. Now, this blog is for you if you want to get familiar with the basic concepts of Postgre SQL. This blog will cover the basic CRUD operation (Create Read Update Delete) in Continue Reading

PostgreSQL OR APACHE CASSANDRA: WHICH ONE IS THE BETTER OPTION

Reading Time: 3 minutes We are living in the 20th century the century of technologies. Because of this, we come across a lot of data in our daily life. So it is important for us to have a database that can help in maintaining a huge amount of data. Now we have many popular databases in the market like PostgreSQL, Cassandra, MySQL MongoDB, and many more. But the question Continue Reading

Exploring HepPlanner for Apache Calcite

Reading Time: 3 minutes In this blog, we will see different ways to manipulate a rel node tree using a Hep planner. A basic understanding of Apache Calcite is necessary for this. Check out the homepage here https://calcite.apache.org/ What is a HepPlanner? It is a rule-based planner to transform a relational expression represented as a tree-like structure. It allows us to specify a condition to identify particular nodes of Continue Reading

Apache Calcite : Adding custom types and functions

Reading Time: 2 minutes Introduction In this blog we will introduce a custom function and type in our SQL . In the end,we want to parse,validate and convert to a relational node for a simple query like“SELECT CAST(my_custom_function(name) as my_custom_type) FROM SAMPLE” . Setting up the basics A sample schema First we need a simple table named Sample : Sample(ID int not null,NAME varchar not null) FrameworkConfig Next we Continue Reading

Reading Excel Data using Fillo API in Selenium

Reading Time: 4 minutes Reading or writing data is one of the most commonly used operations, either fetching values from database tables or fetching values from an excel sheet. it’s will help you to have a Reading Excel Data using Fillo API in Selenium. Fillo API Fillo is a Java API that is used for fetching data from Excel Files. It’s an open source API(created by Codoid) and it’s Continue Reading

Loading JSON data into Snowflake

Reading Time: 4 minutes Have you ever faced any use case or scenario where you’ve to load JSON data into the Snowflake? We better know JSON data is one of the common data format to store and exchange information between systems. JSON is a relatively concise format. If we are implementing a database solution, it is very common that we will come across a system that provides data in Continue Reading

Spark SQL in Delta Lake 0.7.0

Reading Time: 3 minutes Nowadays Delta lake is a buzz word in the Big Data world, especially among the spark developers because it relegates lots of issues found in the Big Data domain. Delta Lake is an open-source storage layer that brings reliability to data lakes. Delta Lake provides ACID transactions, scalable metadata handling, and unifies streaming and batch data processing. It is evolving day by day and adds cool features in its every release. Continue Reading

Optimizations In Spark: For BETTER OR For WORSE

Reading Time: 5 minutes This blog focuses on some of the problems faced while working with the Spark SQL

SQL made easy and secure with Slick

Reading Time: 5 minutes Slick stands for Scala Language-Integrated Connection Kit. It is Functional Relational Mapping (FRM) library for Scala that makes it easy to work with relational databases. Slick can be considered as a replacement of writing SQL queries as Strings with a nicer API for handling connections, fetching results and using a query language, which is integrated more nicely into Scala. You can write your database queries Continue Reading

Installing and Running Presto

Reading Time: 4 minutes Hi Folks ! In my previous blog, I had talked about Getting Introduced with Presto. In today’s blog, I shall be talking about setting up(installing) and running presto. The basic pre-requisites for setting up Presto are: Linux or Mac OS X Java 8, 64-bit Python 2.4+ Installation Download the Presto Tarball from here Unpack the Tarball After unpacking you will see a directory presto-server-0.175 which Continue Reading