Create Oracle Database user with required privileges

Create Oracle Database user with required privileges

Use the following Oracle SQL to create a database user with the typical privileges to connect and work with the database:

CREATE USER dbusername IDENTIFIED BY dbuserpassword;
GRANT CONNECT TO dbusername;


Oracle published results of Customer Support satisfaction survey

Oracle has just published the results of its support satisfaction survey:

And the results are better then expected with a score of 9.11 of 10 (target was 9.01) and an increase from 8.81 of the previous year.

ADWC Performance Benchmark on SSB Schema

The SSB Schema is a generated database schema, that can be used to evaluate the Performance of a Datawarehouse system. This is used to evaluate the performance between ADWC in the cloud and a local Oracle Database which is running within a Virtual Box image on a Laptop. This test is not conducted under test lab conditions and should only provide a rough order of magnitude.

System (Hardware) Baseline

The following instances have been used for the Performance test.

ADWC Trial

based on Hardware specification shown:


This instances is only used as reference, since the sizing (with 2 CPUs) and query performance for some queries based on the same SSB schema are publically available.


instance which only uses 1 CPU

VM Image

with 4GB Ram


and 1 CPU


Based on Laptop Host system (Lenovo Yoga 370 Laptop) with:


The image runs Oracle Database 12c R2 Enterprise Edition, with no specific/tuned configuration (parameters) for Datawarehousing.

System Schema Baseline


Benchmark Performance

For Prologue Queries based on these examples. As mentioned before, for the ADW instance “ADWC Trail”, some queries run times are publically available. These queries have also been executed within the other instances. The runtimes of the ADWC Trail have only been used as reference:


SSB Queries, as outlined here and here (Oracle Documentation).  The below shows the query run-times for the ADWC and local Oracle Database (running as a Virtual Box image). The connection to the ADWC instance is made with the high connection type. When using the medium or low connection type, all queries were showing longer response times (as expected). Thus, the high connection type has been used for conducting the Benchmark.


The Excel file containing the system baseline in terms of number of records, queries and run-time results can be downloaded here.

Below is a galary of the executed SSB Queries on ADW and Virtual Box Oracle Database 12c:

retrieve Table Sizes in Oracle Database

To retrieve the Table Size of a Database Table in Oracle Database, the following SQL code can be used (also without DBA rights to DBA_Segments Data Dictionary Table):


select bytes/1024/1024 MB 
from user_segments 
where segment_name='<Table Name>';


select bytes/1024/1024 MB 
from user_segments 
where segment_name='CUSTOMER';

Use of Date data type in Oracle DV

DV allows to make use of individual elements of any Date or Datetime column. Thus, there is no need to calculate the Year portion of a Date as an additional column. For this to work correclty the data type must also be of Date or Datetime on the Database. When importing a text (csv) file to a Oracle Database, a Date may be defined as a text (varchar) field by default. In order to make use of this DV Feature, the data type should be changed to Date data type when importing e.g. with SQL Developer. Otherwise, the field containing the Date may just be displayed with a blue “A” like any other attribute, but not with the purple “clock” to identify a true Date field.


working with BICC to resolve extract issues

in order to extract data from Oracle Fusion Applications Cloud, BICC needs to be used. This is part of every Fusion Applications Cloud and is typically reachable by extending the URL with /biacm e.g.

When logging into BICC with above example URL, the following screen is shown as starting page:


This is indeed the Review Cloud Extract Configuration page of the top right navigation menu and allows to click on a PVO (Data Store) to further check the configuration of the PVO in terms of associated offerings (OBIA modules) and to be selected attributes (columns) of the PVO (Public View Object).

When starting with a OBIA implementation project, upon the first initial login, the required OBIA modules need to be enabled with the Global Data Store List from the navigation menu:


The offerings (OBIA module) can be selected from the drop down menu and the respective PVOs can be enabled for extract:


This will enable the PVO for use within BICC and the extraction process. In the background these PVOs are created, additionally they will also be associated with the respective OBIA Modules, hence for an OBIA implementation of Human resources, only the required PVOs for HR are extracted from Fusion Cloud.

However, the BICC extracts may fail due to errors which require to execute some tasks within BICC, to fix the Errors as workaround. These tasks include:

  • remove a PVO from an associated offering (OBIA module)
  • deselect columns for a PVO

The following section shows a step-by-step approach to fix these errors within BICC:

remove PVO from offering

A typical reason to remove a PVO from an offering may be due to the following error:

ORA-00942: table or view does not exist 


This error can be resolved by navigating to the start page of BICC or the Review Cloud Extract Configuration link from the navigation menu. The PVO name of the error log can be copied to the Data store field to search. To configure the PVO and the association with the offering the icon right to the name or the name itself can be clicked:


On the page of the PVO click edit:


On the following page, the associated offerings is listed in the (right) selected menu box. The associated offering can be removed from the PVO, to exclude it from the next extract run of the OBIA module (offering):


After removing the associated offering from the PVO, the customization can be saved and on the following screen of PVO summary, the PVO can be closed by clicking “Done” to get back to the Review PVO page.

The above method of customization, will permanently remove the PVO from the offering (OBIA Module). When scheduling a subsequent extract run, the removed PVO will not be part of the offering. This can be validated when creating a new extract (from Manage Extract Schedules) and selecting the offering from the drop-down menu:


The removed PVO will no longer be part of a extract run:


This Data Store List screen can also be used for temporary test of deselecting a PVO from a offering (OBIA module) extract. Typically, the OBIA module is selected from the offering drop-down menu and the enabled for Extract checkbox in the header of the table is clicked to enable all associated PVOs for extract, but an individual PVO can also be deselected for the list, to only affect this specific run (e.g. for temporary testing purposes). Thus, a PVO can be removed from a offering for all subsequent runs or just an specific run (using the Data Store List of the Manage schedule screen).

deselect columns from a PVO

Another common error are missing columns from a PVO e.g. because they are obsolete after an upgrade or not used anymore. The following error is show within the BICC logs:

FscmTopModelAM.PjtProjectPlanAM.ProjectPlanIntegrationPVO – Missing columns – [DeliverableTypesBPEOLastUpdateDate1, ProjPlanLinePEOLastUpdateDate2]





In the latter case, the physical column first needs to be mapped to a attribute (field) of the PVO in order to deselect the attribute from the select list. This is most reliably done by create an SR wth Oracle Support. After the attribute has been identified, the attribute can be deseleted from the PVO by going back to the Review Cloud Extract Configuration screen and searching for the respective PVO. Once the PVO name has been clicked (like in above section) the PVO summary page is shown. From this page the PVO can be edited by clicking “Edit”:


After clicking next to the select column page, the configuration of columns is shown for the PVO:


First, it is recommended to make the highlighted columns visibile to check if an attribute is used for the select list, as primary key or within the incremental filter. The attribute name can then be used to search for the attribute which needs to be deselected:


After deselecting the checkbox e.g. of Select List, the PVO needs to be saved. These changes will then be saved for the next extract run with BICC where the  deselected attribute will not be part of the select statement and thus not be retrieved. Since, this column hasn’t been available and caused the extract issue, the issue has been fixed by also deselecting the attribute from the PVO definition.

OAC: specifying the Connection string within rpd model

The following is a step-by-step guide to determine the Connection string for Oracle Analytics Cloud and Oracle Database Cloud.

When logging into your Oracle Cloud Service using your identity Domain, a similar page is being displayed:


Click on the Action Menu of the Database Cloud Service and select “Open Service Console”.

This Page shows details about your Cloud Services such as OCPUs, Storage or Memory:


click on the Database Service Name to get to the Details Page:


This reveals the Public IP of the Oracle Cloud Database and the required Service Name.

Within the Oracle BI Administration Tool, the connection string needs to be created in the Data Source Name Field:


Using the following format:



After modelling the data and saving the .rpd file, the Model can be uploaded to the OAC Cloud:


Accessing data in Oracle Fusion Applications Cloud for Analytics

The following diagram shows the high-level options to access data within Oracle Fusion Applications (FA) Cloud for Analytical purposes:

Please note: that this diagram is of April 2018 and may not apply in full in future.

For more details on how to connect to Oracle Fusion Applications (FA) Cloud using Visual Analyzer (VA) and DV Desktop (DVD), please reference the previous article.

The details for connecting to FA using ODI (with BI Applications) or BI Data-sync will be described in future articles.