Top 50 Sap Bo Interview Questions You Must Prepare 23.Apr.2024

  •  lookup() : Briefly, It returns single value based on single condition
  •  lookup_ext(): It returns multiple values based on single/multiple condition(s)
  •  lookup_seq(): It returns multiple values based on sequence number

The Data Cleanse transform can generate name match standards and greetings. It can also assign gender codes and prenames such as Mr. and Mrs.

A Parameter is an expression that passes a piece of information to a work flow, data flow or custom function when it is called in a job. A Variable is a symbolic placeholder for values.

Incorrect syntax, Job Server not running, port numbers for Designer and Job Server not matching.

SUPERVISOR offers various types of users of Business Objects products. The user profile determines by default what products a user can use. User profiles can be customized to reflect the needs of the users.

? User (all products but DESIGNER and SUPERVISOR)

? Versatile (configurable)

USERs use the Business Objects end-user products to query, report, and analyze data. They may also use the optional REPORTER and EXPLORER modules for multidimensional analysis. End users can produce documents containing data from one or more data sources.

A versatile user is a customized user who may be given access by a supervisor to any combination of Business Objects products.

These are packages that enhance the ability of Data Cleanse to accurately process various forms of global data by including language-specific reference data and parsing rules.

BOMain. Key is a file which is containing all the information about the repository site .Therefore it will contain the whole address of the repository security domain.

All lookup functions return one row for each row in the source. They differ in how they choose which of several matching rows to return.

  •  Using the auto-correct load option in the target table.
  •  Including the Table Comparison transform in the data flow.
  •  Designing the data flow to completely replace the target table during each execution.
  •  Including a preload SQL statement to execute before the table loads.

Filters are used to get necessary data from a group of data.

Universe is called as the semantic layer between the Database and the one who designs inorder to create the objects and the classes. Universe hence will map to the data in the Database.

Array fetch size indicates the number of rows retrieved in a single request to a source database. The default value is 100@Higher numbers reduce requests, lowering network traffic, and possibly improve performance. The maximum value is 5000

Use of the Case transform is to simplify branch logic in data flows by consolidating case or decision-making logic into one transform. The transform allows you to split a data set into smaller sets based on logical branches.

  •  A job is the smallest unit of work that you can schedule independently for execution.
  •  A work flow defines the decision-making process for executing data flows.
  •  Data flows extract, transform, and load data. Everything having to do with data, including reading sources, transforming data, and loading targets, occurs inside a data flow.

Metrics are a system of parameters .Metrics is a way of quantitative and periodic assessment of a process that is to be measured. Metrics are normally used to track trends, productivity.

  •  Row-by-row select - look up the target table using SQL every time it receives an input row. This option is best if the target table is large.
  •  Cached comparison table — To load the comparison table into memory. This option is best when the table fits into memory and you are comparing the entire target table
  •  Sorted input — To read the comparison table in the order of the primary key column(s) using sequential read.This option improves performance because Data Integrator reads the comparison table only once.Add a query between the source and the Table_Comparison transform. Then, from the query’s input schema, drag the primary key columns into the Order By box of the query.

Name match standards illustrate the multiple ways a name can be represented. They are used in the match process to greatly increase match results.

Business Objects Data Services provides a graphical interface that allows you to easily create jobs that extract data from heterogeneous sources, transform that data to meet the business requirements of your organization, and load the data into a single location.

Business Objects are basically called domain objects. A domain model shows a set of domain objects and explains how they related to each other. Hence a business model also will show a set of domain objects and will show the relationship between them.

A transform enables you to control how datasets change in a dataflow.

Use the USA Regulatory transform if USPS certification and/or additional options such as DPV and Geocode are required. Global Address Cleanse should be utilized when processing multi-country data.

Breaks are used to group the data without changing anything in the format.

Repository Creation will depend on the version. Repository creation will happen after installing the software in 5i/6i versions after installing the software, whereas in the Xi version the repository will be created at the time of installation.

The main solution for improving performance is to make use of the Aggregate tables.

The size of a database will be between 4 and 8 Tera Bytes.

Remove redundant and obsolete objects from the repository tables.

  •  Enhancement Benefit
  •  Determine gender distributions and target
  •  Gender Codes marketing campaigns
  •  Provide fields for improving matching
  •  Match Standards results

Object qualification is just representing what kind of object is that. The three types of object qualifiers are:
@Measure
@Dimension
@Detailed

Join is normally used for linking tables depending upon how the data is requested by the user.

You must define audit points and audit rules when you want to audit a data flow.

A file format is a set of properties describing the structure of a flat file (ASCII). File formats describe the metadata structure. File format objects can describe files in:

  •  Delimited format — Characters such as commas or tabs separate each field.
  •  Fixed width format — The column width is specified by the user.
  •  SAP ERP and R/3 format.

Three.

  •  Database Datastores: provide a simple way to import metadata directly from RDBMS.
  •  Application Datastores: let users easily import metadata from most Enterprise Resource Planning (ERP) systems.
  •  Adapter Datastores: can provide access to an application’s data and metadata or just metadata.

The source of Metrics is by measure objects.

  •  Stage data in an operational datastore, data warehouse, or data mart.
  •  Update staged data in batch or real-time modes.
  •  Create a single environment for developing, testing, and deploying the entire data integration platform.
  •  Manage a single metadata repository to capture the relationships between different extraction and access methods and provide integrated lineage and impact analysis.

The Data Services repository is a set of tables that holds user-created and predefined system objects, source and target metadata, and transformation rules. There are 3 types of repositories.

  •  A local repository
  •  A central repository
  •  A profiler repository

  • Data_Transfer
  •  Date_Generation
  •  Effective_Date
  •  Hierarchy_Flattening
  •  History_Preserving
  •  Key_Generation
  •  Map_CDC_Operation
  •  Pivot Reverse Pivot
  •  Table_Comparison
  •  XML_Pipeline

Constructs a complete hierarchy from parent/child relationships, and then produces a description of the hierarchy in vertically or horizontally flattened format.

  •  Parent Column, Child Column
  •  Parent Attributes, Child Attributes.

Yes. We can have multiple domains

AFD is used to create Dash Boards.

The Data Cleanse transform identifies and isolates specific parts of mixed data, and standardizes your data based on information stored in the parsing dictionary, business rules defined in the rule file, and expressions defined in the pattern file.

The following sections describe ways you can adjust Data Integrator performance

  •  Source-based performance options
  •  Using array fetch size
  •  Caching data
  •  Join ordering
  •  Minimizing extracted data
  •  Target-based performance options
  •  Loading method and rows per commit
  •  Staging tables to speed up auto-correct loads
  •  Job design performance options
  •  Improving throughput
  •  Maximizing the number of pushed-down operations
  •  Minimizing data type conversion
  •  Minimizing local conversion
  •  Improving Informix repository performance

Domain is called as a logical grouping of system tables.

AFDs are normally stored in repository, corporate or personal.

An Embedded Dataflow is a dataflow that is called from inside another dataflow.

conditions are used to get some data based on certain criteria’s or conditions.