Skip to main content

What do you mean by Queued Delta Update Method?

"Queued delta" update method:

With queued delta update mode, the extraction data (for the relevant application) is written in an extraction queue (instead of in the update data as in V3) and can be transferred to the BW delta queues by an update collective run, as previously executed during the V3 update.

After activating this method, up to 10000 document delta/changes to one LUW are cumulated per datasource in the BW delta queues.

If you use this method, it will be necessary to schedule a job to regularly transfer the data to the BW delta queues (by means of so-called "update collective run") by using the same delivered reports as before (RMBWV3); instead, report RSM13005 will not be provided any more since it only processes V3 update entries.

As always, the simplest way to perform scheduling is via the "Job control" function in LBWE.

SAP recommends to schedule this job hourly during normal operation after successful delta initialization, but there is no fixed rule: it depends from peculiarity of every specific situation (business volume, reporting needs and so on).

BENEFITS, BUT
When you need to perform a delta initialization in the OLTP, thanks to the logic of this method, the document postings (relevant for the involved application) can be opened again as soon as the execution of the recompilation run (or runs, if several and running in parallel) ends, that is when setup tables are filled, and a delta init request is posted in BW, because the system is able to collect new document data during the delta init uploading too (with a deeply felt recommendation: remember to avoid update collective run before all delta init requests have been successfully updated in your BW!).

By writing in the extraction queue within the V1 update process (that is more burdened than by using V3), the serialization is ensured by using the enqueue concept, but collective run clearly performs better than the serialized V3 and especially slowing-down due to documents posted in multiple languages does not apply in this method.
On the contrary of direct delta, this process is especially recommended for customers with a high occurrence of documents (more than 10,000 document changes - creation, change or deletion - performed each day for the application in question.

In contrast to the V3 collective run (see OSS Note 409239 ‘Automatically trigger BW loads upon end of V3 updates’ in which this scenario is described), an event handling is possible here, because a definite end for the collective run is identifiable: in fact, when the collective run for an application ends, an event (&MCEX_nn, where nn is the number of the application) is automatically triggered and, thus, it can be used to start a subsequent job.

Besides, don’t omit that queued delta process extraction is independent of success of the V2 update.

REMEMBER TO EMPTY THE QUEUE
The ‘queued delta’ is a good friend, but some care is required to avoid any trouble.

First of all, if you want to take a look to the data of all extract structures queues in Logistic Cockpit, use transaction LBWQ or "Log queue overview" function in LBWE (but here you can see only the queues currently containing extraction data).

In the posting-free phase before a new init run in OLTP, you should always execute (as with the old V3) the update collective run once to make sure to empty the extraction queue from any old delta records (especially if you are already using the extractor) that, otherwise, can cause serious inconsistencies in your data.

Then, if you want to do some change (through LBWE or RSA6) to the extract structures of an application (for which you selected this update method), you have to be absolutely sure that no data is in the extraction queue before executing these changes in the affected systems (and especially before importing these changes in production environment !).

To perform a check when the V3 update is already in use, you can run in the target system the RMCSBWCC check report.

The extraction queues should never contain any data immediately before to:

perform an R/3 or a plug-in upgrade
import an R/3 or a plug-in support packages
Serialized V3 Update(If you set this as your delta mode then data will populated into extraction queue tables)

Comments

Popular posts from this blog

Step by Step guide on BADI with Filter implementation

Step 1 à Before implementing filter objects to a BADI. We need to deactivate all the implementation for that BADI. Go to SE18 (BADI definition) and from there navigate to all BADI implementation to view all the active implementations for that BADI Select the BADI implementations listed on the screen and deactivate the implementation. Once done, return back to the initial screen of SE18 and proceed with next step. Step 2 à go to SE18 – BADI Definition screen and change the existing BADI definition that we created in the earlier tutorial.   Select the FILTER –DEPENDENT check box under the TYPE section. This will open up the FILTER TYPE field for inputs. And do an F1 on the FILTER TYPE field. Click on the F4 on the FILTER TYPE field name and enter some search criteria to find a relevant data element. And click on the continue button   Select the data element from the list to set up your filter and hit enter. Click on the save button and act

Create Data Source in SAP

Data Sources are extraction structure. Data Sources are used to extract data from various source system. In the below steps, we will see how to create a data source using a SAP table VBAK. 1) Execute transaction RSO2 2) Enter data source name to be created 3) Press Create 4) Enter description of the data source 5) Enter the name of table or view on which data source has to be created. Data Source can also be created using a View, Info set or a Function Module. 6) Save the data source 7) Execute RSO2 again and click Change 8) Select VBELN field as selection 9) Save the Data Source 11) To check whether data is getting extracted, execute transaction RSA3 12) Enter the name of data source as shown below and press 'Extraction' 13) An information message will appear stating no. of records fetched 14) Additional button 'Display list' will appear, press button 15) Select one of the data packet and double click to get records fetched 16) Data Source is ready to be used!

Responsibilities of BW Data Architect

Responsibilities of an implementation project... For ex, Lets say If its a fresh implementation of BI or for that matter you are implementing SAP... First and foremost will be your requirements gathering from the client. Depending upon the requirements you will creat a business blueprint of the project which is the entire process from the start to the end of an implementation... After the blue print phase sign off we start off with the realization phase where the actual development happens... In our example after installing the necessary softwares, patches for BI we need to discuss with the end users who are going to use the system for inputs like how they want a report to look like and what are the Key Performance Indicators(KPI) for the reports etc., basically its a question and answer session with the business users... After collecting those informations the development happens in the development servers... After the development comes to an end the same objects are tested in quality