Thursday, June 14, 2012

Essbase Cache Information


Important Notes:
  • Set Retrieval Buffer Size to 50 KB for 32 bit platform and 100 KB for 64 bit platform.
  • Set Retrieval Sort Buffer Size to 50 KB for 32 bit platform and 100 KB on 64 bit platform: Essbase requires a larger retrieval sort buffer size on 64-bit platforms than on 32-bit platforms.


Index cache

The index cache is a buffer in memory that holds index pages. How many index pages are in memory at one time depends upon the amount of memory allocated to the cache.

Data file cache
The data file cache is a buffer in memory that holds compressed data files (.pag files). Essbase allocates memory to the data file cache during data load, calculation, and retrieval operations, as needed. The data file cache is used only when direct I/O is in effect.

Data cache

The data cache is a buffer in memory that holds uncompressed data blocks. Essbase allocates memory to the data cache during data load, calculation, and retrieval operations, as needed.

Calculator cache

The calculator cache is a buffer in memory that Essbase uses to create and track data blocks during calculation operations.

Dynamic calculator cache

The dynamic calculator cache is a buffer in memory that Essbase uses to store all of the blocks needed for a calculation of a Dynamic Calc member in a dense dimension (for example, for a query).

Index size Cache
 

Minimum Value

Default Value

Recommended Value

1024 KB (1048576 bytes)

Buffered I/O: 1024 KB (1048576 bytes)
Direct I/O: 10240 KB (10485760 bytes)
Combined size of all Index Files.
essn.ind files, if possible; as large as possible otherwise. Do not set this cache size higher than the total index size, as no performance improvement results. To determine the total index size, see

Data file Cache
  

Direct I/O: 10240 KB (10485760 bytes)

Direct I/O: 32768 KB (33554432 bytes)

Combined size of all essn.pag files, if possible; otherwise as large as possible.
This cache setting not used if Essbase is set to use buffered I/O.

Data Cache
  

3072 KB (3145728 bytes)

3072 KB (3145728 bytes)

0.125 * the value of data file cache size. Increase value if any of these conditions exist:
  • Many concurrent users are accessing different data blocks.

  • Calculation scripts contain functions on sparse ranges, and the functions require all members of a range to be in memory (for example, when using @RANK and @RANGE).

  • For data load, the number of threads specified by the DLTHREADSWRITE setting is very high and the expanded block size is large.

Calculator Cache
  

Calculator cache

=

Bitmap size in bytes * Number of bitmaps

Bitmap size in bytes

=

Max ((member combinations on the bitmap dimensions/8), 4)

Number of bitmaps

=

Maximum number of dependent parents in the anchoring dimension + 2 constant bitmaps

Minimum Size Specified

Option Selected

625,000 bytes

Option 1 (provides optimal performance)

125,000 bytes

Option 2

2,500 bytes

Option 3

Description

This setting specifies the maximum size Essbase can allocate to each dynamic calculator cache on the server.

Recommended Setting

Recommended setting value = C * S * U.
  • C is the value of the appropriate CALCLOCKBLOCK setting in the essbase.cfg file. (The SET LOCKBLOCK command specifies which CALCLOCKBLOCK setting to use.)

  • S is the size of the largest expanded block across all databases on the machine. To calculate the expanded block size, multiply the number of members (including Dynamic Calc members and dynamic time series members) in each dense dimension together for the number of cells in the block, and then multiply the number of cells by the size of each member cell, 8 bytes.
For example, consider the member count in dense dimensions in Sample Basic:
- 19 (Year, with 12 stored members and 7 Dynamic Calc members, including HTD and QTD)
- 14 (Measures, with 8 stored members and 6 Dynamic Calc members)
- 4 (Scenario, with 2 stored members and 2 Dynamic Calc members)
Note: Label Only members are not counted.
S = 19 * 14 * 4 cells (8 bytes/cell) = 8512 bytes per block
This number is shown in the application log as the logical block size.
  • U is the maximum number of expected concurrent users on the database that has the largest number of concurrent users.
Assigning the value 0 (zero) to DYNCALCACHEMAXSIZE tells Essbase not to use dynamic calculator caches.
By default, the maximum size for this value is 20 MB (20,971,520 bytes).

DYNCALCCACHEWAITFORBLK

Description

If Essbase uses all of the area allocated for a dynamic calculator cache, this setting tells Essbase whether to wait until space becomes available in the cache or to immediately write and calculate the blocks in memory outside the dynamic calculator cache. If the dynamic calculator cache is too small, it is possible for multiple threads to be in queue, each thread waiting to calculate its data blocks.

Recommended Setting

Recommended setting value = FALSE (default value).
Before setting to TRUE, try these alternatives:
  • Add physical memory to the server machine

  • Increase the value of DYNCALCCACHEMAXSIZE, test, and repeat until you verify that you cannot use any more memory for the dynamic calculator cache.

DYNCALCCACHEBLKTIMEOUT

Description

If Essbase is to wait for available space in the dynamic calculator cache, this setting defines how long it waits.

Recommended Setting
Reviewing Dynamic Calculator Cache Usage.
Recommended setting value = WT / B.
  • WT is the maximum tolerable wait time for a query; for example, 5 seconds.

  • B is the total number of logical blocks accessed in the largest query.


To determine the value of B, check the messages in the application log for the largest number of Dyn.Calc.Cache "Big Block Allocs" for a query, as discussed in

DYNCALCCACHEBLKRELEASE

Description

If Essbase has waited the specified time and space is still not available in the dynamic calculator cache, this setting tells Essbase whether to write and calculate the blocks immediately outside the dynamic calculator cache or to create space in the dynamic calculator cache by swapping out blocks and temporarily compressing the swapped blocks in a dynamic calculator cache compressed-block buffer.

Recommended Setting

Recommended setting value = FALSE (default value).
Set to TRUE only if you are experiencing severe memory shortage problems.

DYNCALCCACHECOMPRBLKBUFSIZE

Description

If Essbase has waited the specified wait time and the DYNCALCCACHEBLKRELEASE setting is TRUE, this setting is the size of the dynamic calculator cache compressed-block buffer.

Recommended Setting
essbase.cfg file. The SET LOCKBLOCK command specifies which CALCLOCKBLOCK configuration setting is current.
Recommended setting value = (C * S) / 2.
  • C is the value of the current CALCLOCKBLOCK setting in the

  • S is the size of the largest expanded block across all databases on the machine. Calculate S as described for the DYNCALCCACHEMAXSIZE setting.

This information is found from oracle site. This will be helpful for u all.

Regards,
Suneel Kanthala.

ASO Vs BSO


Agenda
  1. When ASO and BSO Cubes are useful

  2. How ASO Cubes offer new data base types

  3. Performance of ASO Cubes versus BSO Cubes

  4. Differences between ASO and BSO Cubes

  5. When to use ASO Cubes and how users see ASO Cubes
BSO Cubes
  1. Useful for systems that involve complex calculations

  2. BSO Engine cannot handle much sparse data
ASO Cubes
  1. Different method for storage and calculations

  2. Different storage kernel

  3. Useful for databases which are read only

  4. Useful for cases where dimensions are large in number
Eg. Customer analysis, procurement analysis
ASO vs BSO
  1. Load time in ASO is 20% faster than BSO

  2. Calculation time is 20 times faster

  3. Storage space is 10 times smaller

  4. ASO Cubes becomes slower than BSO Cubes for Queries

  5. ASO Cubes can have many more dimensions
End Users
  1. No difference for end users during use with Smart View, Add in, FR Reports
When to use ASO
  1. Database is sparse and has many dimensions. Aggregation is much faster

  2. Calculation of the database is frequent
Storage Kernel of ASO
  1. Multidimensional Index

  2. Indexes the index to facilitate faster aggregation

  3. Unlike BSO Storage is not in Blocks but in optimized aggregation nodes

  4. Loads data at only Level 0

  5. Algorithm selects and stores queries and increased speed

  6. These queries will pre calculate data values and store pre calculate data results in aggregations

    Smart View Timout Issue

    Smart Veiw timeout issue is common problem in the organization. many users installs MSOFFICE in their own systems. some time more data is available when retive or send to data base excel will through the error message as Increase "netRetryCount" and "netRetryInterval"


    For this type of error we need to follow the below steps to overcome the situation.

    Steps as follows:

    1.      Take backup of Window Registry.
    2.      Launch Windows Registry Editor (Regedit). In Windows 7, navigate to C:\Windows and right click on regedit.exe
    3.      Navigate (drill-down) to the following path on the client’s PC Path: [HKEY_CURRENT_USER\Software\Microsoft\Windows\CurrentVersion\Internet Settings]
    4.      Create new DWORD (32-bit) value. Type in “ReceiveTimeout” and press the Enter key.

    5.      Right-Click on the “ReceiveTimeout” registry key.

    6.   Click on “Modify…” from the pop-up list.

    1.      Enter into Value data field: 00dbba00
    2.      Similarly add two new registry keys


     

      

    The Registry Editor for the “Internet Settings” should now have the new keys and values assigned to them.


    If you have any questions regarding this please post comment. I will be answer the questions.

    Regards,
    Suneel Kanthala.

    FDM Application Weblink Error

    Hi,

    If you get error with FDM Url http://localhost/HyperionFDM is not working and giving error message like

    Application Error:

    Description: An exception occured during the execution of the current web request. Please contact administrator to review the stact trace in the event log for more information about the error.

    Solution for this error is:
    1. Stop and start the FDM service. if it works with this option is fine otherwise follow the below steps.
    2.  Run Web logic server-> May be it was stopped when we click on it cmd prompt will get displayed and weblogic server will start in 5-10 minutes.
    3. Go to run and type "inetmngr".
    4. Start the "HyperionFDMAppPool"

    Try these steps The FDM Link will work. if not try to reboot your system this will also solve you problem.

    If you have any questions regarding this please put comments.

    Regards,
    Suneel Kanthala.



    Wednesday, June 13, 2012

    Hyperion Software Products

    • Essbase
    • Hyperion System
    • Hyperion Intelligence (products acquired in 2003 takeover of Brio Technology)
    • Hyperion Enterprise
    • Hyperion Planning
    • Hyperion Strategic Finance
    • Hyperion Financial Management
    • Hyperion Master Data Management (Also reffered to as DRM)
    • Hyperion Financial Reporting
    • Hyperion SmartView
    • Hyperion Financial Data Quality Management (Also referred to as FDM)

    Hyperion Time Line

    • 1981 - IMRS founded by Bob Thomson and Marco Arese, and launches financial and management consolidation software called 'Micro Control' in 1983
    • 1985 - IMRS hires Jim Perakis as CEO; he remains in this position during growth from $1M to almost $300M
    • 1991 - IMRS becomes a public company and launches a Windows-based successor to 'Micro Control' called 'Hyperion'
    • 1992 - Arbor Software ships first version of Essbase Online Analytical processing OLAP software
    • 1995 - IMRS changes name to Hyperion Software Corporation. Arbor becomes a publicly held company
    • 1997 - Arbor acquires Appsource
    • 1998 - Hyperion Software merges with Arbor and the combined company is renamed Hyperion Solutions
    • 1999 - Jeffrey Rodek named as Hyperion Chairman and CEO of Hyperion. Hyperion acquires Sapling Corporation (Enterprise Performance Management applications)
    • 2001 - Godfrey Sullivan is named Hyperion President and COO
    • 2003 - Hyperion acquires Brio Technology and The Alcar Group
    • 2004 - Hyperion names Jeffrey Rodek Executive Chairman; Godfrey Sullivan President and CEO
    • 2005 - Hyperion acquires Razza Solutions (Master data management)
    • 2006 - Hyperion acquires UpStream (Financial Data Quality Management)
    • 2006 - Hyperion acquires Beatware (Data visualization for Web and Mobile Devices)
    • 2007 - Hyperion acquires Decisioneering (Crystal Ball software). Oracle announces agreement to acquire Hyperion for USD 3.3 Billion and bundles Hyperion BI tools into Oracle Business Intelligence Enterprise Edition Plus

    Tuesday, June 12, 2012

    Dataware Housing Concepts


    BASIC DEFINITIONS

    Datawarehousing :
    DWH (Datawarehousing) is a repository of integrated information, specifically structured for Queries and analysis. Data and information are extracted from heterogeneous sources as they are generated. This makes it much easier and more efficient to run queries over data that originally came from different sources.
    "A data warehouse is a subject-oriented, integrated, time-variant, and nonvolatile Collection of data in support of management’s decision making process".
    Subject-oriented – a DW is organized around major subjects; excludes data that is not useful in the decision support process.
    Integrated – a DW is constructed by integrating numerous data sources (relational DB, flat files, legacy systems. DW provides mechanisms for cleaning and standardizing of the data.
    Time-variant – data is stored to provide information from a historical prospective. Every key structure in the data warehouse contains, either implicitly or explicitly, an element of time.
    Nonvolatile – a DW is physically separated from the operational environment. Due to this separation it does not require transaction processing, recovery, and concurrency control mechanisms. It usually requires

    Two operations: initial loading of data and access of data.

    Data Warehouse is an architecture constructed by integrating data from multiple heterogeneous sources to support structured and/or ad hoc queries, analytical reporting and decision making.
    Data Warehousing is a process of constructing and using data warehouses.
      • A Multi-Subject Information Store
      • Typically 100’s of Gigabytes to Terabytes
    Data Mart : It is a collection of subject areas organized for decision support based on the needs of a given department. Ex: sales, marketing etc. the data mart is designed to suit the needs of a department. Data mart is much less granular than the ware house data.

    Data Mart is
      • A Single Subject Data Warehouse
      • Often Departmental or Line of Business Oriented
      • Typically Less Than a 100 Gigabytes
    Differences between DWH & Data Mart : DWH is used on an enterprise level, while data marts are used on a business division / department level. Data warehouses are arranged around the corporate subject areas found in the corporate data model. Data warehouses contain more detail information while most data marts contain more summarized or aggregated data.

    OLTP : OLTP is Online Transaction Processing. This is standard, normalized database structure. OLTP is designed for Transactions, which means that inserts, updates and deletes must be fast.

    OLAP : OLAP is Online Analytical Processing. Read-only, historical, aggregated data.

    Difference between OLTP and OLAP:
    Fact Table :
    It contains the quantitative measures about the business.
    Fact tables that contain aggregated facts are often called summary tables.Dimension Table :
    It is a descriptive data about the facts (business).
    Aggregate tables :
    Aggregate Tables are pre-stored summarized tables. Usage of Aggregates can increase the performance of Queries by several times.
    Conformed dimensions :
    Conformed dimensions are a dimension table shared by fact tables. These tables connect separate star schemas into an enterprise star schema.
    Schema :
    A schema is a collection of database objects, including tables, views, indexes, and synonyms. There are a variety of ways of arranging schema objects in the schema models designed for data warehousing. Most data warehouses use a dimensional model.
    Star Schema :
    Star Schema is a set of tables comprised of a single, central fact table surrounded by de-normalized dimensions. Star schema implement dimensional data structures with de-normalized dimensions
    Snow Flake Schema:
    Snow Flake Schema is a set of tables comprised of a single, central fact table surrounded by normalized dimension hierarchies. Snowflake schema implement dimensional data structures with fully normalized dimensions.
    Queries :
    The DWH contains 2 types of queries. There will be
    • Fixed queries
    • that are clearly defined and well understood, such as regular reports. Ad Hoc Query: Is the starting point for any analysis into a database. The ability to run any query when desired and expect a reasonable response that makes the data warehouse worthwhile and makes the design such a significant challenge. There will also be ad hoc queries that are unpredictable, both in quantity and frequency.
    The end-user access tools are capable of automatically generating the database query that answers any question posted by the user.
    • Canned Queries:
    • are pre-defined queries. Canned queries contain prompts that allow you to customize the query for your specific needs
    Kimball (Bottom up) vs Inmon (Top down) approaches :
    Bottom up:
    Acc. To Ralph Kimball, when you plan to design analytical solutions for an enterprise, try building data marts. When you have 3 or 4 such data marts, you would be having an enterprise wide data warehouse built up automatically without time and effort from exclusively spent on building the EDWH. Because the time required for building a data mart is lesser than for an EDWH.

    Top down:
    Try to build an Enterprise wide Data warehouse first and all the data marts will be the subsets of the EDWH. Acc. To him, independent data marts cannot make up an enterprise data warehouse under any circumstance, but they will remain isolated pieces of information –stove pieces.

    ER Diagram :ER model is a conceptual data model that views the real world as entities and Relationships. A basic component of the model is the Entity-Relationship diagram which is used to visually represent data objects.

    ETL : Extraction, Transformation & Loading. ETL Tools in the market for eg, Informatica, Ascential Data stage, Acta ,Oracle Warehouse Builder(OWB) etc.,

    Staging Area :
    It is the work place where raw data is brought in, cleaned, combined, archived and exported to one or more data marts. The purpose of data staging area is to get data ready for loading into a presentation layer.

    Slowly Changing Dimensions :
    Dimensions are said to be slowly changing dimensions when their attributes remain almost constant, requiring minor alterations.
    Eg Marital status
    Bitmap index, B tree index are the indexing mechanism use for a typical data warehouse.

    OLAP, MOLAP, ROLAP, DOLAP, HOLAP :

    OLAP:
    Online Analytical Processing. OLAP tools in the market eg Business Objects, Brio, Cognos ,Microstrategy , Alphablock, Crystal Reports etc.,

    ROLAP:
    Relationnal OLAP, the users see cubes but under the hood it is pure relational table, Micro-Strategy is a ROLAP product.

    MOLAP:
    Multi dimensionnal OLAP, the users see cubes and under the hood there a big cube, Oracle Express used to be a MOLAP product.

    DOLAP:
    Desktop OLAP, the users see many cubes and under the hood there are many small cubes, Cognos PowerPlay.

    HOLAP:
    Hybrid OLAP, combines MOLAP and ROLAP, Essbase

    Types of Facts:
      1. Additive
        1. Able to add the facts along all the dimensions
        2. Discrete numerical measures eg. Retail sales in $
      2. Nonadditive
        1. Numeric measures that cannot be added across any dimensions
        2. Intensity measure averaged across all dimensions eg. Room temperature
        3. Textual facts - AVOID THEM
      3. Semi Additive
        1. Snapshot, taken at a point in time
        2. Measures of Intensity
        3. Not additive along time dimension eg. Account balance, Inventory balance
        4. Added and divided by number of time period to get a time-average.
    Attributes :
    A field represented by a column within an object (entity). An object may be a table, view or report. An attribute is also associated with an SGML(HTML) tag used to further define the usage.

    Business Activity Monitoring (BAM) :
    BAM is a business solution that is supported by an advanced technical infrastructure that enables rapid insight into new business strategies, the reduction of operating cost by real-time identification of issues and improved process performance.

    Business Intelligence (BI) :Business intelligence is actually an environment in which business users receive data that is reliable,consistent, understandable, easily manipulated and timely. With this data, business users are able to conduct analyses that yield overall understanding of where the business has been, where it is now and where it will be in the near future. Business intelligence serves two main purposes. It monitors the financial and operational health of the organization (reports, alerts, alarms, analysis tools, key performance indicators and dashboards). It also regulates the operation of the organization providing two- way integration with operational systems and information feedback analysis.

    Data Integration :
    Pulling together and reconciling dispersed data for analytic purposes that organizations have maintained in multiple, heterogeneous systems. Data needs to be accessed and extracted, moved and loaded, validated and cleaned, and standardized and transformed.

    Data Mapping :
    The process of assigning a source data element to a target data element.

    Data Mining :
    A technique using software tools geared for the user who typically does not know exactly what he's searching for, but is looking for particular patterns or trends. Data mining is the process of shifting through large amounts of data to produce data content relationships. It can predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. This is also known as data surfing.

    Data Modeling :
    A method used to define and analyze data requirements needed to support the business functions of an enterprise. These data requirements are recorded as a conceptual data model with associated data definitions. Data modeling defines the relationships between data elements and structures.

    Drill Down:
    A method of exploring detailed data that was used in creating a summary level of data. Drill down levels depend on the granularity of the data in the data warehouse.

    Meta Data:
    Meta data is data that expresses the context or relativity of data. Examples of meta data include data element descriptions, data type descriptions, attribute/property descriptions, range/domain descriptions and process/method descriptions. The repository environment encompasses all corporate meta data resources: database catalogs, data dictionaries and navigation services. Meta data includes name, length, valid values and description of a data element. Meta data is stored in a data dictionary and repository. It insulates the data warehouse from changes in the schema of operational systems.

    Normalization:
    The process of reducing a complex data structure into its simplest, most stable structure. In general, the process entails the removal of redundant attributes, keys, and relationships from a conceptual data model.

    Surrogate Key:
    A surrogate key is a single-part, artificially established identifier for an entity. Surrogate key assignment is a special case of derived data - one where the primary key is derived. A common way of deriving surrogate key values is to assign integer values sequentially.

    MOLAP, ROLAP, and HOLAP

    In the OLAP world, there are mainly two different types: Multidimensional OLAP (MOLAP) and Relational OLAP (ROLAP). Hybrid OLAP (HOLAP) refers to technologies that combine MOLAP and ROLAP.

    MOLAP
    This is the more traditional way of OLAP analysis. In MOLAP, data is stored in a multidimensional cube. The storage is not in the relational database, but in proprietary formats.

    Advantages:
    1. Excellent performance: MOLAP cubes are built for fast data retrieval, and is optimal for slicing and dicing operations.
    2. Can perform complex calculations: All calculations have been pre-generated when the cube is created. Hence, complex calculations are not only doable, but they return quickly.
    Disadvantages:
    1. Limited in the amount of data it can handle: Because all calculations are performed when the cube is built, it is not possible to include a large amount of data in the cube itself. This is not to say that the data in the cube cannot be derived from a large amount of data. Indeed, this is possible. But in this case, only summary-level information will be included in the cube itself.
    2. Requires additional investment: Cube technology are often proprietary and do not already exist in the organization. Therefore, to adopt MOLAP technology, chances are additional investments in human and capital resources are needed.
    ROLAPThis methodology relies on manipulating the data stored in the relational database to give the appearance of traditional OLAP's slicing and dicing functionality. In essence, each action of slicing and dicing is equivalent to adding a "WHERE" clause in the SQL statement.


    Advantages:
    • Can handle large amounts of data: The data size limitation of ROLAP technology is the limitation on data size of the underlying relational database. In other words, ROLAP itself places no limitation on data amount.
    • Can leverage functionalities inherent in the relational database: Often, relational database already comes with a host of functionalities. ROLAP technologies, since they sit on top of the relational database, can therefore leverage these functionalities.
    Disadvantages:
    • Performance can be slow: Because each ROLAP report is essentially a SQL query (or multiple SQL queries) in the relational database, the query time can be long if the underlying data size is large.
    • Limited by SQL functionalities: Because ROLAP technology mainly relies on generating SQL statements to query the relational database, and SQL statements do not fit all needs (for example, it is difficult to perform complex calculations using SQL), ROLAP technologies are therefore traditionally limited by what SQL can do. ROLAP vendors have mitigated this risk by building into the tool out-of-the-box complex functions as well as the ability to allow users to define their own functions.
    HOLAP
    HOLAP technologies attempt to combine the advantages of MOLAP and ROLAP. For summary-type information, HOLAP leverages cube technology for faster performance. When detail information is needed, HOLAP can "drill through" from the cube into the underlying relational data.