Like us on Facebook

No of Viewers

Flag Counter

Friday, 6 December 2013

Cognos BI


Why Cognos software?
From business intelligence to financial performance and strategy management to analytics applications, IBM Cognos software can provide what your organization needs to become top-performing and analytics-driven. With products for the individual, workgroup, department, midsize business and large enterprise, Cognos software is designed to help everyone in your organization make the decisions that achieve better business outcomes—for now and in the future. Choose the business intelligence or performance management product you want and be confident that it can grow seamlessly as your needs expand.
Cognos is IBM's business intelligence (BI) and performance management software suite. The software is designed to enable business users without technical knowledge to extract corporate data, analyze it and assemble reports. 
Cognos is composed of nearly three dozen software products. Because Cognos is built on open standards, the software products can be used with relational and multidimensional data sources from multiple vendors, including Microsoft, NCR Teradata, SAP and Oracle. 
The Cognos 8 BI suite, which was first introduced in 2005, is IBM’s leading performance management software. Cognos 8 BI’s capabilities include reporting, analysis, dashboarding and scorecards, delivered through a web-based service-oriented architecture (SOA).
In October 2010, IBM released Cognos 10, which integrates Cogos with other IBM products, including SPSS Predictive Analytics and Lotus Connect. Cognos 10 comes with mobile capability, allowing users to access a complete version of Cognos from mobile devices such as tablets and smartphones.
IBM also offers Cognos Express, which is designed for small to mid-sized companies. The Express edition allows companies to utilize the functions of Cognos 8 BI, while also adding planning, budgeting, and forecasting functions.
IBM’s Cognos products are used by over 23,000 companies around the world, and are consistently ranked in the “leaders” category of Gartner Magic Quadrants for Business Intelligence
Cognos (Cognos Incorporated) was an Ottawa, Ontario-based company making business intelligence (BI) and performance management (PM) software. Founded in 1969, at its peak Cognos employed almost 3,500 people and served more than 23,000 customers in over 135 countries.[1]
Originally Quasar Systems Limited, it adopted the Cognos name in 1982. Cognos is a fragment scissored off the Latin word "cognosco," which means "knowledge from personal experience".[2] January 31, 2008, Cognos was officially acquired by IBM.[3] The Cognos name continues to be applied to IBM's line of business intelligence and performance management products

In January 2010, as part of a reorganization of IBM Software Group, Cognos software and software from recently acquired SPSS were brought together to create the Business Analytics division.
Features
IBM® Cognos® Business Intelligence provides reports, analysis, dashboards and scoreboards to help support the way people think and work when they are trying to understand business performance. You can freely explore information, analyze key facts and quickly collaborate to align decisions with key stakeholders


Reports equip users with the information they need to make fact-based decisions.
Dashboards help users access, interact and personalize content in a way that supports how they make decisions.
Analysis capabilities provide access to information from multiple angles and perspectives so you can view and analyze it to make informed decisions.
Collaboration capabilities include communication tools and social networking to fuel the exchange of ideas during the decision-making process.
Scorecarding capabilities automate the capture, management and monitoring of business metrics so you can compare them with your strategic and operational objectives.
·         Cognos Reporting
·    Reporting is one of the most important part of performance management process. As well as analysis, it guides to identifying three vital, business intelligence questions and, then, finding the answers to them. The questions are: how is a company doing (whether it's on or off track), what is the reason, and - what's the last step - what to do with that? For rationalizing the first stage, that plays an extremely important role - reporting - IBM Cognos solutions contain Query Studio and Report Studio.

     The general, dictionary definition of reporting states that it's data presentation performed to allow target make more efficient decisions. That seems simple. But the real view of reporting depends on perspective it really changes along the person. It might be useful to notice, that more than 95% percent of companies and enterprises have already invested in different reporting solutions and use them in their everyday work. Those solutions have been supposed to support the insight and , thereupon provide better and simpler efficient decision making. Even though reporting is so common and widely used, many of already implemented solutions do not really meet the expectations. Their work (and the results they reach) are far less efficient than it's possible. Therefore, customers are augmented with additional tasks to correct the reports and make them useful enough. As a consequence, reporting itself absorbs too much time, effort, and money than it should, changing the priorities. Necessity is a solution, that supports reporting at its bests. It's thought, that Cognos is one of such solutions.
The choice of best reporting solution depends on a few factors, which defining is indispensable for recognizing the company's requirements.
  • What does reporting exactly mean to a customers?
  • How do they precise their needs?
  • How do they want their reports to look like?
  • For whom do they want to create reports?
  • What is a point of these reports? How are they going to use them?
It's important that both sides - business and IT users - answer the questions above. A comparison of different expectations might help finding the best compromise.

What is wrong with already implemented reporting solutions?


Let's trace an ordinary example. Business unit come across the problem and need to find the "why is that?" answer. Reporting specialists follow their needs to the IT department, where - finally - the report is being prepared basing on owned data. After that, the report is sent back to the orderer. And there start the troubles.
Even though the IT prepares as good report as it's possible, they do not know its reason - the accurate look of the "why?" question. Therefore, the report cannot exactly fit the business unit requirements. Furthermore, all the additional activities, engaging different specialists, multiplied expectations explaining take a lot of time. Time, during which, market conditions change, performance differentiates and situation might significantly turnover. As a consequence, the question itself remains often unanswered and - what's more - the response is wrong and out-of-date.

What has changed?


Even though, business evolves every day, a few change trends are the most significant. The first, and probably the most important thing is business speed. Easier data flow and its higher importance have caused that business users can no longer wait a few days for the report. The time needed for decision making must have been maximally shortened. The second, the meaning of IT resources has been dramatically reduced, due to money limits. Finally, changed the dimension of most businesses - nowadays, former local businesses offer their services in every part of the world, therefore easier data and decision sharing must be provided.

Nowadays, the cooperation of IT and business managers is much closer. Thank to that, the report preparing process lasts shorter and demands significantly less iterative tasks. That results in faster and more accurate reports. Furthermore, the business units do not have to depend on data stored by IT - they have their own resources, that reports might be based on. Then, consumers want reports in different - PDAs, Excel, PDFs - formats. All in all, the need is common. Organizations want a system able to generate reports readable for every consumer they need - independently on consumer's location, time flexibility and format preferences.

Next, across the organization, different types of information and different styles of communication are needed. Other way an employee reports to his corporate officer, and other way he exchanges information with coworkers. As a consequence, an information flow across the organization has a critical meaning.

Business aspires to as high simplicity as possible. What if follows is a need of rapid access to all kinds of information independently on their location and complexity, so employees are able to make responsible decisions immediately. That means that most of decision making process should be possible to be done by business units themselves - without engaging IT. All in all, great diversification of users' requirements comes out from different ways of understanding reporting process.

Reporting styles

 

The basic difference between IBM Cognos and other reporting solutions vendors is a possibility of using different reporting styles. It's come out that many departments across the organization have their own reporting preferences, therefore - to agree them all - the best would be a one tool offering many styles rather than many tools with one style.
  • The first, and the most common reporting style is called managed reporting. A report, created by IT, is developed further to different stakeholders, who then personalize data they got. Using diversified filters and prompts, they can finally receive the information they really wanted. Sales report is its good example - IT share link on a weekly basis, then every customer might customize which products, where sold, what income generated, et cetera, he prefers to compare. Managed reporting might be widely used, due to its personalization options.
  • Second style is preferred usually by operation and line managers, and senior managers, as well. This style reports are being created by professional authors (often from IT). The feature characterizing these reports is their clarity - they're created in a way that allows a customer find what he needs almost immediately. These reports apply to operational database, pointing out the performance. And - the last - they're called transactional and operational reports.
  • Next is a statement report style. Indeed, it's a financial report, but it doesn't necessarily provide any financial intelligence. These reports - created usually by someone from the business units - contain all the useful information - detailed financial data, assets, expenses, liabilities and so on.
  • Production reports contain a number of characteristics and diversified volumes. Made by professional authors, production reports demand fulltime printer readiness. The best example of production reports is Visa statement.
  • Finally, there are also ad hoc reports. Its flexibility of form makes it significantly different from the other styles. Ad hoc reports are prepared only when such a need comes out - they're not daily, weekly, or even monthly prepared. If a customer need - he might prepare an ad hoc report once a year, but also five reports a day. Ad hoc reports are concentrated accurately on finding answer to a concrete question. Then, they may be deleted or sent further, as well. There's no rule. These reports often demand including data from diversified sources. Also, a high formatting discretion is important - user has to be enabled to find the answers he needs as soon as possible, without struggling with complicated formatting. Therefore, IT is very rarely involved in preparing ad hoc reports - that would cost too much time. Thereupon, this kind reports are usually created by business managers for business managers.

How cognos meets these expectations?


What makes it one of the best solutions?
A few things. Once prepared report might be easy shared, and open and read, then modified everywhere. A compatibility with diversified formats. And - what's probably the most convincing - the ability of preparing all styles reports and migrating data among them.

What does it mean in practice?

First of all, once made report might be consumed everywhere - report developer doesn't have to think about the format he saves his results in. Across Cognos, every user can read documents prepared in every format, therefore data transforming problems (and following it delays) stop existing. Cognos supports delivering information, not the documents themselves. Thereupon, the receiver of the report might read it anywhere he wants - independently if it's PDA, Excel, et cetera. It results in real concentration on the information included in the report, not on its form. About rest cares Cognos. What does it result in? For business, it is a significant reduction of business intelligence costs. Not only thinking about money. That's also very important save of a time - preparing one report instead of many in different formats, reduces the time that passes since asking a question till finding a proper answer - that results in more actual data. Furthermore, less people are involved in doing single task.

Second think applies to ad hoc reports. Surely, their making is facilitated by Cognos ad hoc reporting capability, but the best thing starts a while later. Manager finds the answer he was looking for and, then start thinking, what to do with the report he prepared. While he finds it useful for other people in organization, sharing the report is almost trouble-free. An author forwards his report to IT and ask them to turn it into a professional looking and readable for all one. It takes one minute. In opposite to other, much more diversified solutions, which rebuilding the report in, demands almost starting the whole work from the beginning. That once again points out how important a good consolidation and synchronization are Along compatibility with differentiated file formats, come diversified reporting styles. All supported by a one tool. With Cognos, you'll never have to fit your questions to program's abilities. Cognos lets you turn the report in a way that best fits your question's requirements.
Next, the complexity was strongly reduced. Thereupon, users more rarely have to ask IT for help. More and more customers may prepare the reports themselves, what results in time reduction and, also relieves IT departments. Then,
what's connected, data presented in reports is much more timely. The waiting for the report has been maximally reduced, therefore decision making is more efficient than ever before. Moreover, general, but complete reports enable a total end to end view - creating a few separate reports is no longer necessary.

One more thing - let me summarize the information in a word. Business intelligence might be simplified into answering three questions. Two of them - how and why - might be supported by computer. Reporting - presenting processed information and data - applies to the first of them - how is a company doing?
There are a few (mentioned five, but certainly it's possible to find another) styles of reporting. Their choice depends on a customer requirements. The things that diversify reporting styles are their complexity, visibility, generality, accuracy, timeliness, and so on. The more users, the more differentiated expectations, therefore a solution able to agree all of them is a necessity. Ladies and gentlemen, please kindly welcome IBM Cognos 8.


Monday, 18 November 2013

MicroStrategy


What is MicroStrategy?
A Business Intelligence software providing integrated reporting, analysis and monitoring on web and mobile, that helps the Decision Support System of an organization. MicroStrategy is a leading Business Intelligence tool which existing independently by itself (for now).
Technically: A business intelligence tool with a ROLAP engine which could connect to any (or most) database(s). MicroStrategy is an object oriented  model with an attractive and an easy to use graphical user interface for developers and end users.
Basics of MicroStrategy
Microstrategy is an object oriented software, which provides an interface to build reports and dashboards against standard RDMS or cubes. Users will be able to access MicroStrategy reports and dashboards via browsers, mobile (native app) and email in various formats.
Core MicroStrategy Components:
         Intelligence Server.
         Web and Mobile Server.
         MicroStrategy Desktop (client used to connect to MicroStrategy).
         MicroStrategy metadata (database).

MicroStrategy Architect





Intelligence Server: MicroStrategy Intelligent server provides jobs management and analytical processing for all MicroStrategy applications. This acts as a central component connecting the metadata, warehouse, desktop, Web server and Narrow cast Server. Few or main features: Reports Services, OLAP Services, Data Mining, Multi Source connection, Caching, Clustering. Latest version supports installing I-server (different packages) on Windows, Sun Solaris, IBM AIX, HP-UX, Linux.
Web Server: MicroStrategy web server responds to the requests from browsers. Web server interacts with the I-server to extract the necessary information. Can be installed on most of the major web servers and supports most popular browsers. Contact MicroStrategy for certified products.
Narrow cast Server: Narrowcast sever delivers personalized business insight to emails, cell phones, pagers, file servers and print servers extending the reach of Business Intelligence applications. It offers a comprehensive solution for information delivery integrating a subscription portal with a delivery engine.
Metadata: The database repository where definitions of all MicroStrategy objects are stored. Metadata could be hosted on most databases. In simple words, Metadata could be considered as the heart of MicroStrategy environment. Cases where Metadata is directly connected to desktop (client) eliminating I-server is a 2-tier mode connection (refer to the above picture).
Desktop: MicroStrategy Desktop is a client used to interact with the Server.
Types Of Objects (at a high level):
Configuration Objects: Configuration objects are MicroStrategy objects which can be re used in multiple projects and they appear in the system layer.
Schema Objects: The building block of BI. Schema objects are directly mapped to a column or columns in the database.
Public Objects: Objects that generate analytical data and are built on other schema objects or public objects. Also called as application objects.
Types Of Configuration: Objects:
Database Instances, Users, Login ID's, Schedules.
Types Of Schema Objects:
Attributes, Facts, Functions & Operators, Hierarchies, Partition Mappings, Tables & Transformations.
Types Of Public Objects (frequently used):
Consolidation, Custom Groups, Drill Maps, Reports, Documents, Filters, Prompts, Metrics, Templates and Searches
MicroStrategy Roles
Administrator: By default, the role/person will have full access to the environment. In other words this role has full access to all the type of objects mentioned above.
Architect: By default, access to configuration objects is restricted.
Developer: By default, no access to configuration objects, use access to schema objects and full access to public objects
MicroStrategy remains one of the few independent, publicly traded BI software providers in the business intelligence (BI) market. Its primary business analytics competitors include SAP Business Objects, IBM Cognos and Oracle's BI Platform.

MicroStrategy is headquartered in the Washington, D.C. metro area and has worldwide operations across 26 countries, in cities including New York, San Francisco, Paris, Warsaw, Madrid, London, Milan, Dubai, Tokyo, Sydney, and São Paulo. The current CEO and chairman of the board of MicroStrategy is Michael J. Saylor, the company's co-founder. Sanju K. Bansal, the other co-founder, is executive vice president and vice chairman of the board

MicroStrategy is an enterprise business intelligence (BI) application software vendor. The MicroStrategy platform supports interactive dashboards, scorecards, highly formatted reports, ad hoc query, thresholds and alerts, and automated report distribution. Interfaces include web, desktop (for developers) and Microsoft Office integration. MicroStrategy Mobile also supports mobile BI. 

Unlike a traditional multidimensional OLAP (MOLAP) architecture, which supports summary level reporting, MicroStrategy’s relational OLAP (ROLAP) architecture allows users to "drill anywhere" in the entire relational database, all the way down to transactional-level detail.  MicroStrategy has optimizations for all major relational database and data warehouse vendors and can also access multidimensional databases and flat files. While many BI vendors offer full-featured solutions, MicroStrategy’s ROLAP architecture and integrated metadata are key differentiators.

The MicroStrategy platform uses a single common metadata for consistency and streamlined maintenance.  MicroStrategy’s 64-bit architecture supports in-memory analytics with “Intelligent Cubes” (i.e. OLAP reports cached in memory as data sets). Metrics and attributes are created once and used across different types of reports. Changes are made in one place and all related reports
are automatically updated.  Similarly, security permissions are granted in one place, reducing administration costs.

In addition to their suite of development and administrative tools, MicroStrategy provides a software developer kit (SDK) to customize the application and integrate with other applications.  Founded in 1989, MicroStrategy has built its entire integrated product line from the ground up and remains an independent vendor.

Thursday, 10 October 2013

Informatica


Informatica Overview
The Informatica Platform Simplifies Your ETL Processes
Initiate ETL Projects Quickly and Cost-Effectively
Serving as the foundation for all data integration projects, the Informatica Platform lets IT organizations initiate the ETL process from virtually any business system, in any format. As part of the Informatica Platform, Informatica PowerCenter delivers robust yet easy-to-use ETL capabilities that simplify the development and deployment of smaller departmental data marts and data warehouses. In addition, the ETL capabilities facilitate reuse from one project to another.
Enhance ETL with Universal Data Access Capabilities
PowerCenter improves the flexibility of your ETL process with the ability to extract more enterprise data-types than any other technology on the market. Complemented by Informatica PowerExchange and PowerCenter Options, PowerCenter delivers successful ETL initiatives with access to virtually any enterprise data-type, including:
  • Structured, unstructured, and semi-structured data
  • Relational, mainframe, file, and standards-based data
  • Message queue data
  • Automate Most ETL Processes for Fewer Errors and Greater Productivity
PowerCenter makes your ETL developers' jobs easier with cross-functional tools, reusable components, and an enterprise-wide platform that automates many ETL processes. For data warehousing and ETL developers, that means fewer ETL errors and emergency fixes, less risk of rework, faster development time, and greater productivity.
Features
PowerCenter Enterprise forms the foundation for all your data and enterprise integration initiatives—including data governance, data migration, and enterprise data warehousing—setting the standard for high-performance enterprise data integration and quality software. PowerCenter Enterprise scales to support large volumes of disparate data sources and meets demands for security and performance.
Development Agility
Collaborative,
Team-Based Development

PowerCenter provides a variety of graphical development environments designed for data integration developers as well as business users. The shared metadata repository allows groups of users to collaborate on integration projects, enabling rapid iteration cycles that result in significant time savings
Prototype to Production
with a Click
Users can create virtual prototypes of reports and integration jobs without having to move data from its original sources. They'll be able to profile, integrate, and cleanse data on the fly, prototyping integration work in hours instead of weeks. With a few clicks, prototypes can be converted to a physical integration without recoding.
Automated Test Development
Typically, 30 percent of software development is spent on testing code. Data integration projects are no different. Only PowerCenter provides testing tools that automatically generate test cases—saving anywhere from 50 to 80 percent of testing effort while providing significantly more test coverage than manual approaches.

Management Confidence
Reliability and High Availability
Informatica customers count on PowerCenter Enterprise to run their critical business processes. Our high-availability products provide checkpoint recovery so that on the rare occasion that you do have a failure, we pick up the integration job right where it left off.
Traceability and Lineage
Maintaining code and tracking issues can be a daunting challenge when you're hand coding. With an extensible metadata repository that tracks data lineage for you, PowerCenter Enterprise traces the path of data back to the source or its final destination, providing the detailed documentation required by government regulations.
Proactive Monitoring
Only Informatica protects data integration projects with an automated early warning system that alerts your IT team to processes and data quality that differ from the norm as soon as issues arise. PowerCenter Enterprise lets IT monitor workflows, sessions, change control activities, and correlate events across multiple systems
Informatica PowerCenter architecture
Informatica PowerCenter uses a client-server architecture containing several components, described in general terms below, and illustrated in Figure 1. You may find it useful to familiarize yourself with PowerCenter’s architecture before beginning the installation.
For a detailed description of the components that make up PowerCenter, see Chapter 1, “Product Overview,” in Informatica PowerCenter Getting Started.
Informatica PowerCenter contains the following components licensed for use with AX Datasource:
  • Informatica domain – The primary unit for management and administration of services in PowerCenter. Your license agreement restricts you to a single domain.
  • Node – A logical representation of a machine in a domain. The node that hosts the domain is the master gateway for the domain. Your license agreement restricts you to a single node.
  • Informatica Services – A Windows service that starts the Service Manager on a node.
  • Service Manager – Starts and runs the application services on a machine in a domain.
  • Integration Service – Reads workflow information from the PowerCenter repository, and runs sessions and workflows that extract, transform, and load data.
  • Repository Service – Manages connections to the PowerCenter repository.
  • Informatica Administrator – A Web application for managing the Informatica domain, PowerCenter security, and the PowerCenter repository.
  • Informatica domain configuration database – Stores the information (metadata) related to the configuration of the Informatica domain.
  • PowerCenter repository – Stores the information (metadata) required to extract, transform, and load data. Resides in a relational database.
  • PowerCenter Client, which consists of:
    • Designer – Allows you to define sources and targets, and create mappings with transformation instructions, for use in workflows.
    • Workflow Manager – Allows you to create, schedule, and run workflows.
    • Workflow Monitor – Allows you to monitor scheduled and running workflows.
    • Repository Manager – Allows you to administer the PowerCenter repository: assign permissions to users and groups, manage folders, and view PowerCenter repository metadata.

Name changes

The following name changes have occurred

Previous name (in Informatica PowerCenter 8.6.1)New name (in Informatica PowerCenter 9.0.1)
PowerCenter ServerPowerCenter
PowerCenter Administration ConsoleInformatica Administrator
PowerCenter domainInformatica domain

Deployment Flexibility
Get the Connectivity You Deserve
Whether it's structured data in a database, unstructured data like emails or PDF files, social media data in the Cloud, or enterprise applications like SAP or Oracle applications, PowerCenter Enterprise has a high-speed connector that makes data integration quick and easy.
Map Once, Deploy Anywhere
Only Informatica provides a single graphical environment that lets developers create data integration and quality mappings that can be implemented across a variety of technologies. Powered by Vibe™, PowerCenter Enterprise gives you the flexibility to support deployment virtually, on proprietary ETL engines, even on Hadoop, without any recoding.
Meet Your Data Delivery Needs
Not all data is created equal—and neither are your data delivery needs. PowerCenter Enterprise provides a wide variety of technologies for performing data integration, meeting needs from big data and batch to real-time ultra-messaging for high-speed trading.
·       Advanced XML Data Integration Option
The Informatica PowerCenter Advanced XML Data Integration Option enables real-time access to hierarchical data otherwise locked in XML files and messages.
·       Data Integration Analyst Option
Available for Informatica PowerCenter and Informatica Data Services, the Data Integration Analyst option empowers business analysts to perform data integration tasks themselves while IT retains control of the overall data integration process.
·       Data Validation Option
The Informatica PowerCenter Data Validation Option reduces the time and costs of upgrade testing, data integration project testing, and production data auditing and verification by up to 90%—with no programming skills required.
·       Enterprise Grid Option
The Informatica PowerCenter Enterprise Grid Option adds PowerCenter's native data integration grid capabilities, including partitioning and high availability, for more cost-effective performance, dynamic scalability and reliability.
·       High Availability Option
The Informatica PowerCenter High Availability Option minimizes service interruptions during hardware and/or software outages.
·       Metadata Exchange Option
The Informatica PowerCenter Metadata Exchange Options provide access to technical and business metadata from third-party data modeling tools, business intelligence software, and source and target database catalogs.
·       Partitioning Option
The Informatica PowerCenter Partitioning Option helps IT organizations take advantage of parallel data processing in multiprocessor and grid-based hardware environments.
·       Pushdown Optimization Option
The Informatica PowerCenter Pushdown Optimization Option enables data transformation processing, where appropriate, to be pushed down into relational databases or appliances to improve overall performance and throughput.
·       Unstructured Data Option
The Informatica PowerCenter Unstructured Data Option expands PowerCenter's data access capabilities to include unstructured data formats, providing virtually unlimited access to all data formats.

Components under Workflow Manager Tool

Here is a concise idea about the Workflow Manager in Informatica.

1.  Workflow: This is top level object and the entire task (process) for data loading has to be defined under the workflow. It is like a Mapping that integrates different kind of tasks as a Unit.

2.  Task: A task is an individual process to perform a very specific activity during data loading. There are 10 different kinds of tasks that can be grouped under a Workflow:

     2.1 Session:
This is a compulsory task for data loading.
A session is an instance of Mapping Program or in other words a running instance of a mapping is referred as Session. For one Mapping Program we can create one or more Sessions. Generally we require one session for one mapping but for Parallel data loading we may create multiple sessions.

     2.2 Command:
            To execute operating System Commands or Programs. For example: If we need to inform all the users about data loading process, we  can write Shell Script at Operating System and execute them via 'Command Task' just before the session execution.

     2.3 Email:
            To send emails to users using Mail Server (if configured). This job can be done via Command Task also but Email Task is integrated part of Workflow Manager and is much simple compared to Command Task.

     2.4 Decision:
            This task is used to evaluate condition based on other tasks' values to decide next course of actions. It is like IF statement.

     2.5  Control:
            This task is to control the flow of tasks within the Workflow. For example: If we need that control should not reach to specific task (like a Command Task) when a condition fails then we can use Control Task.

     2.6 Event Wait:
                     This task is used to define an event and when the particular event fires (activates) then process continues.

     2.7  Event Raise:
                               This task is used to fire (activate) an event forcefully.

     2.8  Assignment:
This task is used to assign values to Parameters and Variables used within workflow.

     2.9  Timer:
            This task is used to specify time of execution (delay) for a task.

     2.10  Worklet:
This task is to define reusable Workflow. If we need to execute set of tasks again and again under different Workflows then it is better to define them as Worklet and use under different Workflows.
Note:
Three types of tasks (Session, Command and Email) can be defined as Reusable Tasks. Reusable task means: A task is created as an Independent Task and it is used within WorkFlow or Worklet. So if a task is created within workflow or worklet directly then it is non-reusable task.
So if we need a task to be executed once within a Workflow or Worklet then create it as non-reusable task otherwise create them as reusable (independent) task.

Worklet Task also can be defined as Reusable Task via separate Menu Interface.

Friday, 4 October 2013

Teradata

Teradata Overview

Teradata is an enterprise software company that develops and sells a relational database management system (RDBMS) with the same name. In February 2011, Gartner ranked Teradata as one of the leading companies in data warehousing and enterprise analytics. Teradata was a division of the NCR Corporation, which acquired Teradata on February 28, 1991. Teradata's revenues in 2005 were almost $1.5 billion with an operating margin of 21%. On January 8, 2007, NCR announced that it would spin-off Teradata as an independently traded company, and this spin-off was completed October 1 of the same year, with Teradata trading under the NYSE stock symbol TDC.[6]

The Teradata product is referred to as a "data warehouse system" and stores and manages data. The data warehouses use a "shared nothing architecture, which means that each server node has its own memory and processing power. Adding more servers and nodes increases the amount of data that can be stored. The database software sits on top of the servers and spreads the workload among them. Teradata sells applications and software to process different types of data. In 2010, Teradata added text analytics to track unstructured data, such as word processor documents, and semi-structured data, such as spreadsheets.
Teradata's product can be used for business analysis. Data warehouses can track company data, such as sales, customer preferences, product placement, etc.

Teradata is made up of following components –

Processor Chip – The processor is the BRAIN of the Teradata system. It is responsible for all the processing done by the system. All task are done according to the direction of the processor.

Memory – The memory is known as the HAND of the Teradata system. Data is retrieved from the hard drives into memory, where processor manipulates, change or alter the data. Once changes are made in memory, the processor directs the information back to the hard drive for storage.

Hard Drives – This is known as the SPINE of the Teradata system. All the data of the Teradata system is stored in the hard drives. Size of hard drives reflects the size of the Teradata system

Teradata has Linear Scalability
One of the most important asset of Teradata is that it has Linear Scalability. There is no limit on Teradata system. We can grow it to as many times as we want. Any time you want to double the speed of Teradata system, just double the numbers of AMPs and PE. This can be better explained with the help of an example

- Teradata takes every table in the system and spread evenly among different AMPs. Each Amp works on the portion of records which it holds.

- Suppose a EMPLOYEE table has 8 different employee id’s. Now in a 2 AMP system each AMP will hold 4 rows in its DISK to accommodate total 8 rows.

2 AMP SYSTEM

At the time of data retrieval each AMP will work on its DISK and send 4 rows to PE for further processing. If we suppose, one AMP will take 1 microseconds (MS) to retrieve 1 rows, then the time taken to retrieve 4 rows is 4 MS. And as we know that AMPs work in parallel, so both the AMPs will retrieve all 8 records in 4 MS only (4 MS time for each AMP).

Now we double the AMP in our system, and we use total 4 AMP. As Teradata distribute the records evenly among all AMPs, so now each AMP will store 2 records of the table.

4 AMP SYSTEM

Now according to our time scale, the time taken by each AMP for retrieving 2 records is 2MS.
So all 4 AMPs, working parallel, will retrieve the 8 records in 2MS only. Which was previously 4MS for the 2 AMP system.

Hence we double our speed by doubling the number of AMPs in our system.

This is the power of parallelism in Teradata. It is also known as ‘DIVIDE and CONQUER’ theory, according to which we are dividing the work equally and getting the result faster. To achieve the desirable speed we can increase the number of AMPs accordingly.

Partition Primary Index – Advantage and Disadvantage

Advantage of Partition Primary Index –
  • Partitioned Primary Index is one of the unique features of Teradata, which is used for distribution of rows based on different partitions so that they can be retrieved much faster than any other conventional approach.
  • Maximum partitions allowed by Teradata – 65,535 ( suggest if any up gradation )
  • It also reduces the overhead of scanning the complete table (or FTS) thus improving performance.
  • In PPI tables row is hashed normally on the basis of its PI, but actual storage of row in AMP will take place only in its respective partition. It means rows are sorted first on the basis of there partition column and then inside that partition they are sorted by there row hash.
  • Usually PPI’s are defined on a table in order to increase query efficiency by avoiding full table scans without the overhead and maintenance costs of secondary indexes.
  • Deletes on the PPI table is much faster.
  • For range based queries we can effectively remove SI and use PPI, thus saving overhead of SI subtable.

Disadvantage of Partition Primary Index –
  • PPI rows are 2 bytes are longer so it will use more PERM space.
  • In case we have defined SI on PPI table then as usual size of SI sub table will also increase by 2 bytes for each referencing rowed
  • A PI access can be degraded if the partition column is not part of the PI. For e.g. if query specifying a PI value but no value for the PPI column must look in each partition for that table, hence loosing the advantage of using PI in where clause.
  • When we are doing joins to non-partitioned tables with the PPI table then that join may be degraded. If one of the tables is partitioned and other one is non-partitioned then sliding window merger join will take place.
  • The PI can’t be defined UNIQUE when the portioning columns are not the part of PI.
Technology and product
Teradata is a massively parallel processing system running, a shared nothing architecture. Its technology consists of hardware, software, database, and consulting. The system moves data to a data warehouse where it can be recalled and analyzed.
The systems can be used as back-up for one another during downtime, and in normal operation balance the work load across themselves.
In 2009, Forrester Research issued a report, "The Forrester Wave: Enterprise Data Warehouse Platform," by James Kobielus, rating Teradata the industry's number one enterprise data warehouse platform in the "Current Offering" category.
Marketing research company Gartner Group placed Teradata in the "leaders quadrant" in its 2009, 2010, and 2012 reports, "Magic Quadrant for Data Warehouse Database Management Systems".
Teradata is the most popular data warehouse DBMS in the DB-Engines database ranking.
In 2010, Teradata was listed in Fortune’s annual list of Most Admired Companies

Active enterprise data warehouse

Teradata Active Enterprise Data Warehouse is the platform that runs the Teradata Database, with added data management tools and data mining software.
The data warehouse differentiates between “hot and cold” data – meaning that the warehouse puts data that is not often used in a slower storage section. As of October 2010, Teradata uses Xeon 5600 processors for the server nodes.
Teradata Database 13.10 was announced in 2010 as the company’s database software for storing and processing data.
Teradata Database 14 was sold as the upgrade to 13.10 in 2011 and runs multiple data warehouse workloads at the same time. It includes column-store analyses.
Teradata Integrated Analytics is a set of tools for data analysis that resides inside the data warehouse

Backup, archive, and restore

BAR is Teradata’s backup and recovery system.
The Teradata Disaster Recovery Solution is automation and tools for data recovery and archiving. Customer data can be stored in an offsite recovery center.

Platform  family

Teradata Platform Family is a set of products that include the Teradata Data Warehouse, Database, and a set of analytic tools. The platform family is marketed as a smaller and less expensive than the other Teradata solutions