SAP Support Info


ROLE OF EP ADMINISTRATOR:


 i have posted  links to some common activities that a portal administrator might need to do with the tools.


Some common tasks.

This is a list of any child pages with the label ep_admintasks.
 Connecting to the J2EE Engine with the Visual Administrator — You don't remember the port that needs to be specified to connect to the J2EE Engine with the Visual Administrator.
 Implementing a Federated Portal Network on SAP NetWeaver Portal 7 
 To Reset Administrator Password through Config Tool — It may sometimes happen that a user enters the wrong password a number of times (by default, 3) and if that user is the administrator, then the administrator user is locked. This prevents all sorts of things happening. To unlock the administator, use the following process


ROLE OF A BASIS CONSULTANT IN BW PROJECT
On-going Basis Administration support of SAP R/3 systems.
• Customizing and troubleshooting and day-to-day support.
• SAP security through the profile Generator and Management of Users and Authorizations.
• Logon load balancing
• Dat Archiving
• Configuration of operation modes.
• Performance Tuning and Monitoring,
• Lock Management,
• Update Processing,
• View Logs, Short Dumps and Trace Files
• Oracle Database & Table space maintenance
• Workload Analysis
• Performing Client Copy functions
• Batch/Background Jobs Management.
• Support Package administration.
• Backup Management
Creating Transport Landscape (And Packages etc..)
Creating RFC Destinations and Remote users.
Maintain Roles and Authorizations.
Applying the OSS Notes or Support packages
Maintain the Database like Table Spaces etc...
Monitoring the jobs of Bw .Some back groound jobs willl be running.
Handling of connection related issues.
Helping in tRFC / QRFC related issues
Archiving Idocs etc
Archiving base tables in the BW System






















SAP BI Production Support Issues




Production Support Errors :1) Invalid characters while loading: When you are loading data then you may get some special characters like @#$%...e.t.c.then BW will throw an error like Invalid characters then you need to go through this RSKC transaction and enter all the Invalid chars and execute. It will store this data in RSALLOWEDCHAR table. Then reload the data. You won't get any error because now these are eligible chars done by RSKC.

2) IDOC Or TRFC Error: We can see the following error at “Status” Screen:Sending packages from OLTP to BW lead to errorsDiagnosisNo IDocs could be sent to the SAP BW using RFC.System responseThere are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of the SAP BW.Further analysis:Check the TRFC log.You can get to this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".Removing errors:If the TRFC is incorrect, check whether the source system is completely connected to the SAP BW. Check especially the authorizations of the background user in the source system.Action to be taken:If Source System connection is OK Reload the Data.

3)PROCESSING IS OVERDUE FOR PROCESSED IDOCsDiagnosis IDocs were found in the ALE inbox for Source System that is not updated. Processing is overdue. Error correction: Attempt to process the IDocs manually. You can process the IDocs manually using the Wizard or by selecting the IDocs with incorrect status and processing them manually. Analysis:After looking at all the above error messages we find that the IDocs are found in the ALE inbox for Source System that are not Updated.Action to be taken:We can process the IDocs manually via RSMO -> Header Tab -> Click on Process manually.

4) LOCK NOT SET FOR LOADING MASTER DATA ( TEXT / ATTRIBUE / HIERARCHY )Diagnosis User ALEREMOTE is preventing you from loading texts to characteristic 0COSTCENTER . The lock was set by a master data loading process with therequest number. System response For reasons of consistency, the system cannot allow the update to continue, and it has terminated the process. Procedure Wait until the process that is causing the lock is complete. You can call transaction SM12 to display a list of the locks. If a process terminates, the locks that have been set by this process are reset automatically. Analysis:After looking at all the above error messages we find that the user is “Locked”. Action to be taken:Wait for sometime & try reloading the Master Data manually from Info-package at RSA1.

5) Flat File Loading ErrorDetail Error MessageDiagnosis Data records were marked as incorrect in the PSA. System response The data package was not updated.Procedure Correct the incorrect data records in the data package (for example by manually editing them in PSA maintenance). You can find the error message for each record in the PSA by double-clicking on the record status.Analysis:After looking at all the above error messages we find that the PSA contains incorrect record.Action to be taken:To resolve this issue there are two methods:-i) We can rectify the data at the source system & then load the data.ii) We can correct the incorrect record in the PSA & then upload the data into the data target from here.

6) Object requested is currently locked by user ALEREMOTEDetail Error Message.DiagnosisAn error occurred in BI while processing the data. The error is documented in an error message.Object requested is currently locked by user ALEREMOTEProcedureLook in the lock table to establish which user or transaction is using the requested lock (Tools -> Administration -> Monitor -> Lock entries). Analysis:After looking at all the above error messages we find that the Object is “Locked. This must have happened since there might be some other back ground process runningAction to Be taken : Delete the error request. Wait for some time and Repeat the chain.

Idocs between R3 and BW while extraction
1)When BW executes an infopackage for data extraction the BW system sends a Request IDoc ( RSRQST ) to the ALE inbox of the source system.Information bundled in Request IDoc (RSRQST) is :
Request Id ( REQUEST )
Request Date ( REQDATE )
Request Time (REQTIME)
Info-source (ISOURCE)
Update mode (UPDMODE )
2)The source system acknowledges the receipt of this IDoc by sending an Info IDoc (RSINFO) back to BW system.The status is 0 if it is ok or 5 for a failure.
3)Once the source system receives the request IDoc successfully, it processes it according to the information in the request. This request starts the extraction process in the source system (typically a batch job with a naming convention that begins with BI_REQ). The request IDoc status now becomes 53 (application document posted). This status means the system cannot process the IDoc further.
4)The source system confirms the start of the extraction job by the source system to BW by sending another info IDoc (RSINFO) with status = 1
5)Transactional Remote Function Calls (tRFCs) extract and transfer the data to BW in data packages. Another info IDoc (RSINFO) with status = 2 sends information to BW about the data package number and number of records transferred
6)At the conclusion of the data extraction process (i.e., when all the data records are extracted and transferred to BW), an info IDoc (RSINFO) with status = 9 is sent to BW, which confirms the extraction process.


Function module to make yellow request to RED

Use SE37, to execute the function module RSBM_GUI_CHANGE_USTATE.From the next screen, for I_REQUID enter that request ID and execute.From the next screen, select 'Status Erroneous' radiobutton and continue.This Function Module, change the status of request from Green / Yellow to RED. 

What will happend if a request in Green is deleted?

Deleting green request is no harm. if you are loading via psa, you can go to tab 'reconstruction' and select the request and 'insert/reconstruct' to have them back.But,
For example you will need to repeat this delta load from the source system. If you delete the green request then you will not get these delta records from the source system.

Explanation :
when the request is green, the source system gets the message that the data sent was loaded successfully, so the next time the load (delta) is triggered, new records are sent.
If for some reason you need to repeat the same delta load from the source, then making the request red sends the message that the load was not successful, so do not discard these delta records.Delta queue in r/3 will keep until the next upload successfully performed in bw. The same records are then extracted into BW in the next requested delta load. 

PSA reverse posting

When the data is loaded to BW Infocube and then it is compresed then u can't delete the data pertaining to this request in the cube . For this what u can do is if the request is loaded via PSA then u can click on the Request Reverse Posting option on the MOnitor screen of the particular Request . This will reverse the sign of the keyfigures loaded into the InfoCube for that particular request only ,so that it will make overall keyfigyure value in the cube for this particular request to 0


Reverse posting to be done by system @ Monitoring --> Scheduler --> Reverse posting --> Immediate & Save.
This will nullify the before request values by sending reverse values.
This can be done only if the loaded data is still present in PSA. 

Unable to Cancel Job in SM37 (R3)

Question from sdn :
We got a situation :Unable to Kill the Background job (BI_REQ*).It shows "ACTIVE" and been running for more than 19,000 secs (2 days..) .When this job started ,the PROCESS ID was : 244277..But after a day,this PROCESS ID : 244277 has been allocated to some other R3 Program(Monitored in SM50).But the Background Job is still "ACTIVE".Tried killing the Job in SM37--->Cancel Active Job or Delete the Job) but ,when tried to "Cancel" the Job,we get a message : The job is NOT ACTIVE !!!. Then when tried to "DELETE" the job ,we get a message " Job is still ACTIVE"......!!!!
----->Since there is no Work Process in SM50,what is the alternative solution to Kill this Job..........???
(In BW ,there is no Update of data records,with the monitor showing

Answere :
Go to SM37 >> select the job >> Click on the JOB tab in the top >> Click on Check status............
Then it status will show cancel...........
Actually sometimes it happens job gets cancelled but staus shows active....then we need to do this..........actually job is already cancelled......

Attribute delta loading "duplicate record found"

Master data attribute loading error: "x duplicate record found. y recordings used in table z "

Fix:
Option in InfoPackage "Ignore Double Data Records"
OR
Make "Change Run" tool run for every upload.
OR
In start routine of Transfer or Update rules make a sort of record of the package and then delete adiacent duplicates. 

Error loading master data - Data record 1 ('AB031005823') : Version 'AB031005823' is not valid

Problem
Created a flat file datasource for uploading master data.Data loaded fine upto PSA.Once the DTP which runs the transformation is scheduled, its ends up in error as below:

Solution
After refering to many links on sdn, i found that since the data is from an external file,the data will not be matching the SAP internal format. So it shud be followed that we mark "External" format option in the datasource ( in this case for Material ) and apply the conversion routine MATN1 as shown in the picture below:

Once the above changes are done, the load was successful.
Knowledge from SDN forums
Conversion takes place when converting the contents of a screen field from display format to SAP-internal format and vice versa and when outputting with the ABAP statement WRITE, depending on the data type of the field.
Check the info here:
http://help.sap.com/saphelp_nw04/helpdata/en/2b/e9a20d3347b340946c32331c96a64e/content.htm
http://help.sap.com/saphelp_nw04/helpdata/en/07/6de91f463a9b47b1fedb5be18699e7/content.htm
This fm ( MATN1) will add leading ZEROS to the material number because when u query on MAKT with MATNR as just 123 you wll not be getting any values, so u should use this conversion exit to add leading zeros. 



When is reconstruction allowed?

1. When a request is deleted in a ODS/Cube, will it be available under reconstruction.
Ans :Yes it will be available under reconstruction tab, only if the processing is through PSA Note: This function is particularly useful if you are loading deltas, that is, data that you cannot request again from the source system
2. Should the request be turned red before it is deleted from the target so as to enable reconstruction
Ans :To enable reconstruction you may not need to make the request red, but to enable repeat of last delta you have to make the request red before you delete it.
3. If the request is deleted with its status green, does the request get deleted from reconstruction tab too
Ans :No, it wont get deleted from reconstruction tab
4. Does the behaviour of reconstruction and deletion differ when the target is differnet. ODS and Cube
Ans :Yes


How to Debugg Update and transfer Rules
1.Go to the Monitor.
2. Select 'Details' tab.
3. Click the 'Processing'
4. Right click any Data Package.
5. select 'simulate update'
6. Tick the check boxes ' Activate debugging in transfer rules' and 'Activate debugging in update rules'.
7. Click 'Perform simulation'.


Error loading master data - Data record 1 ('AB031005823') : Version 'AB031005823' is not valid
ProblemCreated a flat file datasource for uploading master data.Data loaded fine upto PSA.Once the DTP which runs the transformation is scheduled, its ends up in error as below:


SolutionAfter refering to many links on sdn, i found that since the data is from an external file,the data will not be matching the SAP internal format. So it shud be followed that we mark "External" format option in the datasource ( in this case for Material ) and apply the conversion routine MATN1 as shown in the picture below

:Once the above changes are done, the load was successful.Knowledge from SDN forumsConversion takes place when converting the contents of a screen field from display format to SAP-internal format and vice versa and when outputting with the ABAP statement WRITE, depending on the data type of the field.

Check the info:http://help.sap.com/saphelp_nw04/helpdata/en/2b/e9a20d3347b340946c32331c96a64e/content.htmhttp://help.sap.com/saphelp_nw04/helpdata/en/07/6de91f463a9b47b1fedb5be18699e7/content.htmThis fm ( MATN1) will add leading ZEROS to the material number because when u query on MAKT with MATNR as just 123 you wll not be getting any values, so u should use this conversion exit to add leading zeros.’

Function module to make yellow request to RED
Use SE37, to execute the function module RSBM_GUI_CHANGE_USTATE.From the next screen, for I_REQUID enter that request ID and execute.From the next screen, select 'Status Erroneous' radiobutton and continue.This Function Module, change the status of request from Green / Yellow to RED.

What will happend if a request in Green is deleted?
Deleting green request is no harm. if you are loading via psa, you can go to tab 'reconstruction' and select the request and 'insert/reconstruct' to have them back.But,For example you will need to repeat this delta load from the source system. If you delete the green request then you will not get these delta records from the source system.Explanation :when the request is green, the source system gets the message that the data sent was loaded successfully, so the next time the load (delta) is triggered, new records are sent.If for some reason you need to repeat the same delta load from the source, then making the request red sends the message that the load was not successful, so do not discard these delta records.Delta queue in r/3 will keep until the next upload successfully performed in bw. The same records are then extracted into BW in the next requested delta load.

Appearence of Values for charecterstic input help screen
Which settings can I make for the input help and where can I maintain these settings?In general, the following settings are relevant and can be made for the input help for characteristics:Display: Determines the display of the characteristic values with the following options "Key", "Text", "Key and text" and "Text and key".Text type: If there are different text types (short, medium and long text), this determines which text type is to be used to display the text.Attributes: You can determine for the input help which attributes of the characteristic are displayed initially. When you have a large number of attributes for the characteristic, it makes sense to display only a selected number of attributes. You can also determine the display sequence of the attributes.F4 read mode: Determines in which mode the input help obtains its characteristic values. This includes the modes "Values from the master data table (M)", "Values from the InfoProvider (D)" and "Values from the Query Navigation (Q)".

Note that you can set a read mode, on the one hand, for the input help for query execution (for example, in the BEx Analyzer or in the BEX Web) and, on the other hand, for the input help for the query definition (in the BEx Query Designer). You can make these settings in InfoObject maintenance using transaction RSD1 in the context of the characteristic itself, in the InfoProvider-specific characteristic settings using transaction RSDCUBE in the context of the characteristic within an InfoProvider or in the BEx Query Designer in the context of the characteristic within a query. Note that not all the settings can be maintained in all the contexts. The following table shows where certain settings can be made:

Setting RSD1 RSDCUBE BExQueryDesigner
Display X X X
Text type X X X
Attributes X - -
Read mode -
Query execution X X X -
Query definition X - -
Note that the respective input helps in the BEx Web as well as in the BEx Tools enable you to make these settings again after executing the input help.


When do I use the settings from InfoObject maintenance (transaction RSD1) for the characteristic for the input help?

The settings that are made in InfoObject maintenance are active in the context of the characteristic and may be overwritten at higher levels if required. At present, the InfoProvider-specific settings and the BEx Query Designer belong to the higher levels. If the characteristic settings are not explicitly overwritten in the higher levels, the characteristic settings from InfoObject maintenance are active.When do I use the settings from the InfoProvider-specific characteristic settings (transaction RSDCUBE) for the input help?You can make InfoProvider-specific characteristic settings in transaction RSDCUBE -> context menu for a characteristic -> InfoProvider-specific properties.These settings for the characteristic are active in the context of the characteristic within an InfoProvider and may be overwritten in higher levels if required. At present, only the BEx Query Designer belongs to the higher levels. If the characteristic settings are not explicitly overwritten in the higher levels and settings are made in the InfoProvider-specific settings, these are then active. Note that the settings are thus overwritten in InfoObject maintenance.When do I use the settings in the BEx Query Designer for characteristics for the input help?In the BEx Query Designer, you can make the input help-relevant settings when you go to the tab pages "Display" and "Advanced" in the "Properties" area for the characteristic if this is selected.These settings for the characteristic are active in the context of the characteristic within a query and cannot be overwritten in higher levels at present. If the settings are not made explicitly, the settings that are made in the lower levels take effect.

How to supress messages generated by BW Queries
Standard Solution :
You might be aware of a standard solution. In transaction RSRT, select your query and click on the "message" button. Now you can determine which messages for the chosen query are not to be shown to the user in the front-end.

Custom Solution:
Only selected messages can be suppressed using the standard solution. However, there's a clever way you can implement your own solution... and you don't need to modify the system for it!All messages are collected using function RRMS_MESSAGE_HANDLING. So all you have to do is implement an enhancement at the start of this function module. Now it's easy. Code your own logic to check the input parameters like the message class and number and skip the remainder of the processing logic if you don't want this message to show up in the front-end.

FUNCTION rrms_message_handling.
StartENHANCEMENT 1 Z_CHECK_BIA.
* Filter BIA Message
if i_class = 'RSD_TREX' and i_type = 'W' and i_number = '136'*
just testing it.*
exitend if.
ENHANCEMENT
End
IMPORTING
------------
----------
----
EXCEPTIONS
Dummy ..

How can I display attributes for the characteristic in the input help?
Attributes for the characteristic can be displayed in the respective filter dialogs in the BEx Java Web or in the BEx Tools using the settings dialogs for the characteristic. Refer to the related application documentation for more details.In addition, you can determine the initial visibility and the display sequence of the attributes in InfoObject maintenance on the tab page "Attributes" -> "Detail" -> column "Sequence F4". Attributes marked with "0" are not displayed initially in the input help.

Why do the settings for the input help from the BEx Query Designer and from the InfoProvider-specific characteristic settings not take effect on the variable screen?
On the variable screen, you use input helps for selecting characteristic values for variables that are based on characteristics. Since variables from different queries and from potentially different InfoProviders can be merged on the variable screen, you cannot clearly determine which settings should be used from the different queries or InfoProviders. For this reason, you can use only the settings on the variable screen that were made in InfoObject maintenance.

Why do the read mode settings for the characteristic and the provider-specific read mode settings not take effect during the execution of a query in the BEx Analyzer?

The query read mode settings always take effect in the BEx Analyzer during the execution of a query. If no setting was made in the BEx Query Designer, then default read mode Q (query) is used.

How can I change settings for the input help on the variable screen in the BEx Java Web?

In the BEx Java Web, at present, you can make settings for the input help only using InfoObject maintenance. You can no longer change these settings subsequently on the variable screen.

Selective Deletion in Process Chain
The standard procedure :
Use Program RSDRD_DELETE_FACTS
1. Create a variant which is stored in the table RSDRBATCHPARA for the selection to be deleted from a data target.
2. Execute the generated program.
Observations:
The generated program executes will delete the data from data target based on the given selections. The program also removes the variant created for this selective deletion in the RSDRBATCHPARA table. So this generated program wont delete on the second execution.

If we want to use this program for scheduling in the process chain we can comment the step where the program remove the deletion of the generated variant.

Eg:REPORT ZSEL_DELETE_QM_C10 .
TYPE-POOLS: RSDRD, RSDQ, RSSG.
DATA:
L_UID TYPE RSSG_UNI_IDC25,
L_T_MSG TYPE RS_T_MSG,
L_THX_SEL TYPE RSDRD_THX_SEL
L_UID = 'D2OP7A6385IJRCKQCQP6W4CCW'.
IMPORT I_THX_SEL TO L_THX_SEL
FROM DATABASE RSDRBATCHPARA(DE) ID L_UID.
* DELETE FROM DATABASE RSDRBATCHPARA(DE) ID L_UID.CALL FUNCTION 'RSDRD_SEL_DELETION'
EXPORTING
I_DATATARGET = '0QM_C10'
I_THX_SEL =
L_THX_SELI_AUTHORITY_CHECK = 'X'
I_THRESHOLD = '1.0000E-01'
I_MODE = 'C'
I_NO_LOGGING = ''
I_PARALLEL_DEGREE = 1
I_NO_COMMIT = ''
I_WORK_ON_PARTITIONS = ''
I_REBUILD_BIA = ''
I_WRITE_APPLICATION_LOG = 'X'
CHANGING
C_T_MSG =
L_T_MSG.export l_t_msg to memory id sy-repid.
UPDATE RSDRBATCHREP
SET DELETEABLE = 'X'
WHERE REPID = 'ZSEL_DELETE_QM_C10'.


ABAP program to find prev request in cube and delete
There will be cases when we cannot use the SAP built-in settings to delete previous request..The logic to determine previous request may be so customised, a requirement.In such cases you can write a ABAP program which calculates previous request basing our own defined logic.Following are the tables used : RSICCONT ---(list of all requests in any particular cube)RSSELDONE ----- ( has got Reqnumb, source , target , selection infoobject , selections ..etc)Following is one example code. Logic is to select request based on selection conditions used in the infopackage:


TCURF, TCURR and TCURX
TCURF is always used in reference to Exchange rate.( in case of currency translation ).For example, Say we want to convert fig's from FROM curr to TO curr at Daily avg rate (M) and we have an exchange rate as 2,642.34. Factors for this currency combination for M in TCURF are say 100,000:1.Now the effective exchange rate becomes 0.02642.
Question ( taken from sdn ):can't we have an exchange rate of 0.02642 and not at all use the factors from TCURF table?.I suppose we have to still maintain factors as 1:1 in TCURF table if we are using exchange rate as 0.02642. am I right?. But why is this so?. Can't I get rid off TCURF.What is the use of TCURF co-existing with TCURR.Answer :Normally it's used to allow you a greater precision in calaculationsie 0.00011 with no factors gives a different result to0.00111 with factor of 10:1So basing on the above answer, TCURF allows greater precision in calculations.Its factor shud be considered before considering exchange rate

.-------------------------------------------------------------------------------------TCURRTCURR table is generally used while we create currency conversion types.The currency conversion types will refer to the entries in TCURR defined against each currency ( with time reference) and get the exchange rate factor from source currency to target currency.

-------------------------------------------------------------------------------------
TCURXTCURX
table is used to exactly define the correct number of decimal places for any currency. It shows effect in the BEx report output.
-------------------------------------------------------------------------------------

How to define F4 Order Help for infoobject for reporting
Open attributes tab of infoobject definition.In that you will observe column for F4 order help against each attribute of that infoobject like below :
This field defines whether and where the attribute should appear in the value help.Valid values:• 00: The attribute does not appear in the value help.•
01: The attribute appears at the first position (to the left) in the value help.•
02: The attribute appears at the second position in the valuehelp.•
03: ......• Altogether, only 40 fields are permitted in the input help. In addition to the attributes, the characteristic itsel, its texts, and the compounded characteristics are also generated in the input help. The total number of these fields cannot exceed 40.
So accordingly , the inofobjects are changed> Suppose if say for infobject 0vendor, if in case 0country ( which is an attribute of 0vendor) is not be shown in the F4 help of 0vendor , then mark 0 against the attribtue 0country in the infoobject definition of 0vendor.

Dimension Size Vs Fact Size
The current size of all dimensions can be monitored in relation to fact table by t-code se38 running report SAP_INFOCUBE_DESIGNS.Also,we can test the infocube design by RSRV tests.It gives out the dimension to fact ratio.

The ratio of a dimension should be less than 10% of the fact table.In the report,Dimension table looks like /BI[C/O]/D[xxx]
Fact table looks like /BI[C/0]/[E/F][xxx]
Use T-CODE LISTSCHEMA to show the different tables associated with a cube.

When a dimension grows very large in relation to the fact table, db optimizer can't choose efficient path to the data because the guideline of each dimension having less than 10 percent of the fact table's records has been violated.

The condition of having large data growth in a dimension is called degenerative dimension.To fix, move the characteristics to different dimensions. But can only be done when no data in the InfoCube.

Note : In case if you have requirement to include item level details in the cube, then may be the Dim to Fact size will obviously be more which you cant help it.But you can make the item charecterstic to be in a line item dimension in that case.Line item dimension is a dimension having only one charecterstic in it.In this case, Since there is only one charecterstic in the dimension, the fact table entry can directly link with the SID of the charecterstic without using any DIMid (Dimid in dimension table usually connects the SID of the charecterstic with the fact) .Since link happens by ignoring dimension table ( not in real sense ) , this will have faster query performance.

BW Main tables
Extractor related tables: ROOSOURCE - On source system R/3 server, filter by: OBJVERS = 'A'
Data source / DS type / delta type/ extract method (table or function module) / etc
RODELTAM - Delta type lookup table.
ROIDOCPRMS - Control parameters for data transfer from the source system, result of "SBIW - General setting - Maintain Control Parameters for Data Transfer" on OLTP system.
maxsize: Maximum size of a data packet in kilo bytes
STATFRQU: Frequency with which status Idocs are sent
MAXPROCS: Maximum number of parallel processes for data transfer
MAXLINES: Maximum Number of Lines in a DataPacketMAXDPAKS: Maximum Number of Data Packages in a Delta RequestSLOGSYS: Source system.

Query related tables:
RSZELTDIR: filter by: OBJVERS = 'A', DEFTP: REP - query, CKF - Calculated key figureReporting component elements, query, variable, structure, formula, etc
RSZELTTXT: Similar to RSZELTDIR. Texts of reporting component elementsTo get a list of query elements built on that cube:RSZELTXREF: filter by: OBJVERS = 'A', INFOCUBE= [cubename]
To get all queries of a cube:RSRREPDIR: filter by: OBJVERS = 'A', INFOCUBE= [cubename]To get query change status (version, last changed by, owner) of a cube:RSZCOMPDIR: OBJVERS = 'A' .

Workbooks related tables:
RSRWBINDEX List of binary large objects (Excel workbooks)
RSRWBINDEXT Titles of binary objects (Excel workbooks)
RSRWBSTORE Storage for binary large objects (Excel workbooks)
RSRWBTEMPLATE Assignment of Excel workbooks as personal templatesRSRWORKBOOK 'Where-used list' for reports in workbooks.

Web templates tables:
RSZWOBJ Storage of the Web Objects
RSZWOBJTXT Texts for Templates/Items/Views
RSZWOBJXREF Structure of the BW Objects in a TemplateRSZWTEMPLATE Header Table for BW HTML Templates.

Data target loading/status tables:
rsreqdone, " Request-Data
rsseldone, " Selection for current Request
rsiccont, " Request posted to which InfoCube
rsdcube, " Directory of InfoCubes / InfoProvider
rsdcubet, " Texts for the InfoCubes
rsmonfact, " Fact table monitor
rsdodso, " Directory of all ODS Objects
rsdodsot, " Texts of ODS Objectssscrfields. " Fields on selection screens

Tables holding charactoristics:
RSDCHABAS: fields
OBJVERS -> A = active; M=modified; D=delivered
(business content characteristics that have only D version and no A version means not activated yet)TXTTABFL -> = x -> has text
ATTRIBFL -> = x -> has attribute
RODCHABAS: with fields TXTSHFL,TXTMDFL,TXTLGFL,ATTRIBFL
RSREQICODS. requests in ods
RSMONICTAB: all requestsTransfer Structures live in PSAPODSD
/BIC/B0000174000 Trannsfer Structure
Master Data lives in PSAPSTABD
/BIC/HXXXXXXX Hierarchy:XXXXXXXX
/BIC/IXXXXXXX SID Structure of hierarchies:
/BIC/JXXXXXXX Hierarchy intervals
/BIC/KXXXXXXX Conversion of hierarchy nodes - SID:
/BIC/PXXXXXXX Master data (time-independent):
/BIC/SXXXXXXX Master data IDs:
/BIC/TXXXXXXX Texts: Char./BIC/XXXXXXXX Attribute SID table:

Master Data views
/BIC/MXXXXXXX master data tables:
/BIC/RXXXXXXX View SIDs and values:
/BIC/ZXXXXXXX View hierarchy SIDs and nodes:InfoCube Names in PSAPDIMD
/BIC/Dcube_name1 Dimension 1....../BIC/Dcube_nameA Dimension 10
/BIC/Dcube_nameB Dimension 11
/BIC/Dcube_nameC Dimension 12
/BIC/Dcube_nameD Dimension 13
/BIC/Dcube_nameP Data Packet
/BIC/Dcube_nameT Time/BIC/Dcube_nameU Unit
PSAPFACTD
/BIC/Ecube_name Fact Table (inactive)/BIC/Fcube_name Fact table (active)

ODS Table names (PSAPODSD)
BW3.5/BIC/AXXXXXXX00 ODS object XXXXXXX : Actve records
/BIC/AXXXXXXX40 ODS object XXXXXXX : New records
/BIC/AXXXXXXX50 ODS object XXXXXXX : Change log

Previously:
/BIC/AXXXXXXX00 ODS object XXXXXXX : Actve records
/BIC/AXXXXXXX10 ODS object XXXXXXX : New records

T-code tables:
tstc -- table of transaction code, text and program name
tstct - t-code text .

1What is tickets? And example?
The typical tickets in a production Support work could be:
1. Loading any of the missing master data attributes/texts.
2. Create ADHOC hierarchies.
3. Validating the data in Cubes/ODS.
4. If any of the loads runs into errors then resolve it.
5. Add/remove fields in any of the master data/ODS/Cube.
6. Data source Enhancement.
7. Create ADHOC reports.
1. Loading any of the missing master data attributes/texts - This would be done by scheduling the info packages for the attributes/texts mentioned by the client.
2. Create ADHOC hierarchies. - Create hierarchies in RSA1 for the info-object.
3. Validating the data in Cubes/ODS. - By using the Validation reports or by comparing BW data with R/3.
4. If any of the loads runs into errors then resolve it. - Analyze the error and take suitable action.
5. Add/remove fields in any of the master data/ODS/Cube. - Depends upon the requirement
6. Data source Enhancement.
7. Create ADHOC reports. - Create some new reports based on the requirement of client.
Tickets are the tracking tool by which the user will track the work which we do. It can be a change requests or data loads or whatever. They will of types critical or moderate. Critical can be (Need to solve in 1 day or half a day) depends on the client. After solving the ticket will be closed by informing the client that the issue is solved. Tickets are raised at the time of support project these may be any issues, problems.....etc. If the support person faces any issues then he will ask/request to operator to raise a ticket. Operator will raise a ticket and assign it to the respective person. Critical means it is most complicated issues ....depends how you measure this...hope it helps. The concept of Ticket varies from contract to contract in between companies. Generally Ticket raised by the client can be considered based on the priority. Like High Priority, Low priority and so on. If a ticket is of high priority it has to be resolved ASAP. If the ticket is of low priority it must be considered only after attending to high priority tickets.
Checklists for a support project of BPS - To start the checklist:
1) Info Cubes / ODS / data targets 2) planning areas 3) planning levels 4) planning packages 5) planning functions 6) planning layouts 7) global planning sequences 8) profiles 9) list of reports 10) process chains 11) enhancements in update routines 12) any ABAP programs to be run and their logic 13) major bps dev issues 14) major bps production support issues and resolution .

What are the tools to download tickets from client? Are there any standard tools or it depends upon company or client...?Yes there are some tools for that. We use Hpopenview. Depends on client what they use. You are right. There are so many tools available and as you said some clients will develop their own tools using JAVA, ASP and other software. Some clients use just Lotus Notes. Generally 'Vantive' is used for tracking user requests and tickets.
It has a vantive ticket ID, field for description of problem, severity for the business, priority for the user, group assigned etc.
Different technical groups will have different group ID's.
User talks to Level 1 helpdesk and they raise ticket.
If they can solve issue for the issue, fine...else helpdesk assigns ticket to the Level 2 technical group.
Ticket status keeps changing from open, working, resolved, on hold, back from hold, closed etc. The way we handle the tickets vary depending on the client. Some companies use SAP CS to handle the tickets; we have been using Vantage to handle the tickets. The ticket is handled with a change request, when you get the ticket you will have the priority level with which it is to be handled. It comes with a ticket id and all. It's totally a client specific tool. The common features here can be - A ticket Id, - Priority, - Consultant ID/Name, - User ID/Name, - Date of Post, - Resolving Time etc.
There ideally is also a knowledge repository to search for a similar problem and solutions given if it had occurred earlier. You can also have training manuals (with screen shots) for simple transactions like viewing a query, saving a workbook etc so that such queried can be addressed by using them.
When the problem is logged on to you as a consultant, you need to analyze the problem, check if you have a similar problem occurred earlier and use ready solutions, find out the exact server on which this has occurred etc.
You have to solve the problem (assuming you will have access to the dev system) and post the solution and ask the user to test after the preliminary testing from your side. Get it transported to production once tested and posts it as closed i.e. the ticket has to be closed.

3.What is User Authorizations in SAP BW?
Authorizations are very important, for example you don't want the important financial report to all the users. so, you can have authorization in Object level if you want to keep the authorization for specific in object for this you have to check the Object as an authorization relevant in RSD1 and RSSM tcodes. Similarly you set up the authorization for certain users by giving that users certain auth. in PFCG tcode. Similarly you create a role and include the tcodes; BEx reports etc into the role and assign this role to the userid.


Transporting a change to the structure of an ODS object

Adding fields is a change to the structure of the ODS. When you import the transport request, the existing tables of the ODS have to be
converted to reflect the change. After adding those fields to the underlying ODS tables (Alter table...) the fields have to be
filled with initial values. If the ODS is large then these update statements could have very long execution times.
 When you have large amount of data in the table (ODS), the following is a work around to get the new columns added to the ODS/DSO:
1. Make copy of the existing ODS (ODS1) to ODS2
2. Add new fields to ODS2
3. Load the data from ODS1 to ODS2 using datamart (use ODS2 for further
   purposes).
If you have some reporting queries defined on ODS1 then you need to do
the following:
1. Make copy of the existing ODS (ODS1) to ODS2.
2. Load data from ODS1 to ODS2 using datamart.
3. Delete Data from ODS1.
4. Add new fields to ODS1.
5. Reload the data from ODS2 to ODS1 using datamart.

Performance issue during DSO request activation






Here are some general tips on to improve the activation performance of DSO requests:
- DSO activation can also be slow if the batch tables are large as these are run through for object activations. So as a starter, please
clean down the batch system with the usual jobs (RSBTCDEL2, SM65, etc); your Basis team will be aware of these & should run these for you.
- ensure that statistics for the DSO are up to date.
- if you are not reporting on the DSO, the activation of SIDs is not required (this will take up some considerable time in activation); Often the logs show that the activation job takes almost all the time to schedule the RSBATCH_EXECUTE_PROZESS as job BIBCTL_*. RSBATCH_EXECUTE_PROCESS is for scheduling and executing the SID-generation process. If you don't need the relevant DSO for reporting & you don't have queries on it, you can delete the reporting flag in the DSO maintenance. This would be a good way to speed this process up significantly. Check under 'Settings' in the DSO maintenance whether you have flagged the option "SID Generation upon activation".
- by making some adjustments to the DataStore Object parameters in tcode RSODSO_SETTINGS you should be able to accelerate the request.
You can adjust this for all, or specific, DSOs. See SAP Note 1118205 - RSODSO_SETTINGS Maintain runtime parameter of DSO.
- read the SAP Note 1392715 - DSO req. activation:collective perf. problem note


How to Configure alert emails for Process Chain errors?



Would like to share this information regarding email alerts whenerrors occur in Process Chains.Step by step process to follow:
1. go to ALRTCATDEF
2. select classification "Process Chains"
3. click on display/change
4. double-click on "error in a process of a process chain"
5. click on "fixed recipients"
6. write in your username (you have to maintain your e-mail address in SU01 for that to work)
7. save
8. go to your process chain (rspc)
9. go into edit mode
10. from the menu, select Process chain -> attributes -> alerting
11. check "send alerts if errors occur"
  • Steps 1 to 7 should be done once to configure.
  • Follow steps 8 to 10 for the process chains for which alert emails are required. 
List of OSS Notes for SAP NetWeaver 7.0 BI Query Performance






Increase query performance  through Information Broadcasting of key queries by filling  the OLAP cache or the MDX cache. In case of large number of users accessing a query or have query that access high volume of data.
Note 1026944 - New Cache mode for BI 7.0 without directory
Note 1105139 - Filling the OLAP cache with "contains pattern" in BI 7.0
Note 1115731 - Fill MDX cache: "Portal theme & 1 does not fit output format
Note 1108347 - OLAP cache GETWA_NOT_ASSIGNED in CL_RSR_CACHE_QUERY_CUBE_MEM
Note 1101187 - The OLAP Tunnel
Note 1117348 - Subsequent correction to Note 1101187 (OLAP Tunnel)
Note 1144702 - Memory release additional corrections to Note 1101187
Note 989793 - Deactivating the read statistics for OLAP Cache BI 7.0
Note 859456 - Increasing performance using the cache
Note 1114164 - Termination in the cache due to insufficient shared memory
Note 1093755 - Statistics counter of OLAP cache not increased
The lock table is a table in the main memory of the enqueue server that records the current locks in the system.
Note 1090847 - RSR_CACHE: Wait time too long during locking for main memory
Note 1107434 - OLAP cache consultation: Long wait times due to locks
Increase query performance by making sure that queries are regenerated in production using RSRT after changes to statistics, consistency changes, or aggregates.
Note 1022589 - The query settings  RSRT - Query Mass Maintenance
Note 1136163 - The query settings in RSRT - Attributes
Improve SAP NetWeaver Business Intelligence reporting performance using statistical content. Important SAP Notes related to statistical content set up guide for configuring the SAP NetWeaver BI Administration Cockpit and guide for activating technical content.
Note 834280 - Installing technical BI Content after upgrade
Note 965386 - Activating the technical content for the BI Admin Cockpit
Note 934848 - Collective Note : (FAQ) BI Administration Cockpit
Improve SAP NetWeaver 7.0 BI Query Performance
Note 1025307 - Composite Note for NW2004s performance: Reporting
Note 990399 - Performance problems when documents are used in web reports
Note 764394 - Guide for performance improvement of Web Applications
Note 1090514 - Improving Performance of queries with two structures & Hier
Note 1024554 - Improving Performance in queries in SAPLRSEC_CHECKS
Note 1113195 - Improving Performance when there are several data providers
Note 1083175 - IP - Preliminary analysis of performance problems
Note 1056259 - Collective Note - Performance of BI Planning
Note 1089831 - Enterprise search: Performance when searching for queries
Note 1145255 - Query Performance Check corrections - Hotfix SP5 - Part 1 
Analyse performance of BI queries and templates in the BI Java Runtime
Note 1048691 - NW 7.0 BI Web Applications: Performance Problem Analysis
Note 1061240 - Slow web browser due to JavaScript virus scan

Delta Administration Tables and Important Fields





ROOSPRMSC 
RLOGSYS: target system
SLOGSYS: source system
INITRNR: init request
DELTARNR: last delta request
INITSTATE: if X than the delta update is active
ROOSPRMSF  selection criteria for init request 
FIELDNM
RSLOW
RSHIGH

RSSDLINIT
LOGSYS: source system
RNR: Init request number
INITSTATE: = X -> Delta update is active
LAST DELTA RNR: last delta update

RSSDLINITSEL selection criteria for init request
IOBJNM
FIELDNAME
RSLOW
RSHIGH

RSSDLINITDEL all deleted initalisations
Source System Authorizations




The RFC user on BW, receiving data from OLTP system, should have
profiles:
- S_BI-WHM_RFC
- S_RS_ALL
The RFC user on OLTP, receiving request from BW system, should have
profiles:
- S_BI-WX_RFC
- S_IDOC_ALL
- B_ALE_ALL

Control Parameters for extraction

In every SAP source system, part of the ROIDOCPRMS table controls the data transfer from the source system to the BW system. The table
contains the following information:
- MAXSIZE - Maximum size of a data package in KB
- STATFRQU - Number of packets that are transferred before statistical information is sent
- MAXLINES - Maximum number of records sent in one data package
- MAXPROCS - Maximum number of dialog work processes for each upload request used to send the data to the BW system
Max. proc are the number of parallel processes for data transfer.The more parallel processes that can run the better the performance of
your system will be. However, you must bear in mind the number of available dialog processes here (as opposed to BTCs).
Take note:
- In a resource constrained system, reduce the data package size.
- In larger systems (with adequate CPU and memory) increase the package size to speed up collection. However, take care so as not to impact the communication process and unnecessarily  hold up the work processes


Data Transfer Process and Error handling process



Data Transfer Process and Error handling process
Introduction about Data Transfer Process
  1. You use the data transfer process (DTP) to transfer data within BI from a persistent object to another object in accordance with certain transformations and filters. In this respect, it replaces the data mart interface and the Info Package. As of SAP Net Weaver 2004s, the Info Package only loads data to the entry layer of BI (PSA).
  2. The data transfer process makes the transfer processes in the data warehousing layer more transparent. Optimized parallel processing improves the performance of the transfer process .You can use the data transfer process to separate delta processes for different targets and you can use filter options between the persistent objects on various levels. For example, you can use filters between a Data Store object and an Info Cube.
  3. Data transfer processes are used for standard data transfer, for real-time data acquisition and for accessing data directly. 

     
    Interesting Benefits of New Data Transfer Process
  1. Loading data from one layer to others except Info sources.
  2. Separation of delta mechanism for different data targets.
  3. Enhanced filtering in dataflow.
  4. Improved transparency of staging processes across data warehouse layers.
  5. Improved performance : optimized parallelization
  6. Enhanced error handling in the form of error stack
  7. Enables real-time data acquisition.  
                                   
       
Most important advantage in Data Transfer Process
  1. Delta logic can be separately handled for separate data targets
  2. Example for separation for delta logic
  3. Delta logic is a part of DTP
  4. One Source PSA
  5. Two targets : One DSO keeping daily data and other one keeping weekly data 

Five process for handling errors in DTP 
Process#1 Enhanced Filtering, Debugging and error handling options
Process # 2 -Handling Data Records With Errors
  1. Using the error handling settings on the Update tab page in the data transfer process, when data is transferred from a DTP source to a DTP target, you can specify how the system is to react if errors occur in the data records. 
  2. These settings were previously made in the Info Package. When using data transfer processes, Info Packages write to the PSA only. Error handling settings are therefore no longer made in the Info Package, but in the data transfer process 
Process # 3 -Error Handling Features
  1. Possibility to choose in the scheduler to
  2. Abort process when errors occur
  3. Process the correct records but do not allow reporting on them
  4. Process the correct records and allow reporting on them
  5. Number of wrong records which lead to a wrong request
  6. Invalid records can be written into an error stack
  7. Keys should be defined for error stack to enable the error handling of data store object
  8. Temporary data storage can be switched on/off for each sub step of the loading process
  9. Invalid records can be updated into data records after their correction
Process # 4 - Error Stack
  1. Stores erroneous records
  2. Keeps the right sequence of records à for consistent data store handling.
  3. Key of error stack defines which data should be detained from the update after the erroneous data record.
  4. After Correction, Error-DTP updates data from error stack to data target.
Note: Once the request in the source object is deleted, the related data records in error stack area automatically deleted.



    1.  Key of Error Stack = Semantic Groups.
    2.  Subset of the key of the target object.


      1. Max. 16 fields
      2. Defines which data should be detained from the update of erroneous data record (for data store object)
      3. The bigger the key, the fewer records will be written to the error stack 
        Process # 5 - Temporary Data Storage
  1. In order to analyze the data at various stages you can activate the temporary storage in the DTP
  2. This allows you to determine the reasons of error 
    I enjoyed sharing my knowledge. Hope the above information is useful and provide me your feed back. Thank you. 


1 comment:

  1. It is amazing and wonderful to visit your site.Thanks for sharing this information,this is useful to me...
    http://chennaitraining.in/sap-mm-training-in-chennai/
    http://chennaitraining.in/sap-mss-training-in-chennai/
    http://chennaitraining.in/sap-pi-training-in-chennai/
    http://chennaitraining.in/sap-pm-training-in-chennai/
    http://chennaitraining.in/sap-pp-training-in-chennai/
    http://chennaitraining.in/sap-ps-training-in-chennai/
    http://chennaitraining.in/sap-qm-training-in-chennai/

    ReplyDelete