Skip to main content

SQL Server 70-461 Exams fundamental - Part 1 (Tips from Querying from Microsoft SQL Server Training kit by Itzik-Ben-Gan)

Hello Everybody!!! This time, it is about Microsoft Exam.
When I decided to take up Microsoft certification (MCSA) on Business Intelligence 2012 (I was planning from 3 years but was never that serious to write, Lol!! But now I am ready), I came up with training kit by Itzik-Ben-Gan. This book was very helpful for understanding the fundamentals of T-SQL and its architectures.

However, for my preparation standpoint view, I am summarizing the tips and points from the book which I believe, going through main points, will help lot of them to get small tips instead of reading entire book before day of exam. So let’s start with SQL….!!!

Standard SQL is based on the relational model, which is a mathematical model for data management and manipulations.
The word relational is not about 2 tables related but it is from mathematical concept of Set Relation from “Set theory”.

Relation in the relational model is called a “Table”. (T-SQL is based on multiset theory than on set theory).
For a table to be in Relational format, set of rules are applicable but if not followed, those tables are not really called relational or in other words, the so called relational table deviates from mathematical Set theory. The examples of such deviations are below.

For a table to be in Relational format, set of rules are applicable but if not followed, those tables are not really called relational or in other words, the so called relational table deviates from mathematical Set theory. The examples of such deviations are below.

Ø  Set theory say, A Set cannot have duplicates. This is not true when it comes to Table. T-SQL on querying a set of records, by default will not remove duplicates.
Ø  Adding DISTINCT which remove duplicates will make the table relational.
Ø  Set theory doesn’t believe in order of tuples (sets like a,b,c,z,y) arranged. T-SQL on enforcing the rules like ORDER BY clause, will make the table not relational or deviates from relational model.
(Note: SQL Server doesn’t track the order of records been saved into its memory. It randomly picks the location and no guarantee that SQL Server would save the data on particular address. The SQL Server stores the rows in a volatile fashion).
Ø  According to Relational theory, all the attributes should have names. It deviates in T-SQL where we still can write queries like,
Select EmpID, Firstname+ ‘  ‘ +LastName From Emp.
Where 2nd column resulting is derived column from Firstname and LastName but have no doesn’t have any name.
This can be overcome by;
 Select EmpID, Firstname+ ‘  ‘ +LastName as Name From Emp.
Ø  Same name of column cannot be a output if it is relational. Example ;
Select E1.ID  , E2.ID from … can result in column ID from 2 tables but this is not going to form relational table under relational model. Each column should be identify distinctly.
Ø  Set theory stands for 2 valued- logic (True or a False) but T-SQL is a three valued-logic system. It follows or predicates TRUE or FALSE OR UNKNOWN (NULL- Meaning missing value).
Ex: Select * From EMP Where Ename IS NULL.

Comments

Popular posts from this blog

BIG Data, Hadoop – Chapter 2 - Data Life Cycle

Data Life Cycle The data life cycle is pictorial defined as show below:     As we see, in our current system, we capture/ Extract our data, then we store it and later we process for reporting and analytics. But in case of big data, the problem lies in storing and then processing it faster. Hence Hadoop takes this portion, where it stores the data in effective format (Hadoop distributed File System) and also process using its engine (Map Reduce Engine). Since Map Reduce engine or Hadoop engine need data on HDFS format to process, We have favorable tools available in market to do this operation. As an example, Scoop is a tool which converts RDBMS to HDFS. Likewise we have SAP BOD to convert sap system data to HDFS.

SSIS: The Value Was Too Large To Fit In The Output Column

I had a SSIS package where I was calling a stored procedure in OLEDB Source and it was returning a “The Value Was Too Large to Fit in the Output Column” error. Well, My Datatype in OLEDB source was matching with my OLEDB Destination table. However, when I googled, we got solutions like to increase the output of OLEDB Source using Advanced Editor option . I was not at all comfortable with their solution as my source, destination and my intermediate transformation all are having same length and data type and I don’t want to change. Then I found that I was missing SET NOCOUNT ON option was missing in Stored Procedure. Once I added it, my data flow task ran successfully. 

How to Copy or Move Multiple Files from One Folder to Another Folder using Talend

Hello all, In this Post, I will explain how to move Multiple Files from One Folder (Say Source) to Other folder (Say Destination). This Post will also helps you to understand How to Declare Variable and Use it. To Declare a variable, We are go to use Contexts option in repository. Lets say we have two .txt files in Path D:/Source/ . My Requirement is to move the files from Source Folder ( D:/Source/ ) to Destination Folder ( D:/Dest/ ). Step 1: Open a New job Step 2: Now right click and Create a New Contexts from Repository. Give some Name and give Next. Step 3: Now Fill in the Source Directory Details where the loop on files should happen as shown in the snippet and give finish. Step 4: Now Context is created and The values will be changing based on each file in Folder. Step 5: Click and Drag the context from Repository to Context Job Window below the Job Designer. Step 6: If we Expand the Contexts, We can find the variable SourcePath is holdi...