Mauro decided to give me another challenge- run SQLd360 against the AWR Warehouse and let him know how it does straight “out of the box”. It’s a simpler installation process than SQLTXPLAIN, you simply unzip and run, (no installation of any packages…)
I chose a SQL_ID from one of the source databases loaded and ran it for a 31 day history:
I finally have a moment to look into my “hacked” SQLT XPRECT/XTRACT runs and see if the changes I made were enough to run properly with the AWR Warehouse.
The answer is yes and no… and maybe “it depends”…
If you are a resident of California, you can become a member of the San Jose Public Library and gain access to Safari Books Online and Books24x7. At last count, there were 445 Oracle titles available through Safari Books Online(read more)
P.O.I.S.E.D. is the acronym I invented for the performance tuning method described by Chris Lawson in The Art and Science of Oracle Performance Tuning; it stands for Problem, Observation, Interpretation, Solution, Execution, and Documentation. The Observation and Interpretation phases feed each other. The Interpretation phase concludes with a determination of root cause. If we are investigating poor performance a single SQL statement, the tool of choice for the observation phase is SQLT for all but the simplest cases.(read more)
Last year I promised myself I would do more conferences & presenting and that it would include more events further afield, such as in Europe. I can’t say I managed it in 2014 (Liverpool for the UKOUG Tech14 did not count as a foreign country for me, even if I found a couple of the locals hard to understand) but 2015 is proving more successful. I attended the OUG Ireland conference 2 weeks ago, for my first trip to that country, and I learnt recently that I have papers accepted for Harmony 2015. This conference is a joint event between the Oracle user group of Finland, the Oracle user group of Estonia and the Latvian Oracle user group.
Es sind noch ein paar Plätze frei - sollten Sie also Lust und Zeit haben, nach Berlin zu kommen und exklusives Wissen (nicht nur) über das Parallel Execution Feature der Oracle Datenbank zu erfahren, würde ich mich sehr freuen, Sie dort mit den anderen Teilnehmern begrüßen zu dürfen, um gemeinsam mit Ihnen eine gute und produktive Zeit zu verbringen!
Bei Interesse wenden Sie sich bitte an die Ansprechpartner der DOAG, die im Link angegeben sind - dort finden Sie auch eine genauere Beschreibung des Seminars.
Trivadis, the company I work for, just opened a new branch in Denmark (the press release, in German, is available here). On the 14th of April (09:00-15:00) takes place, in the Oracle’s office in Ballerup, the opening event. Two colleagues of mine and I will be there to deliver the following presentations:
Big Data versus Conventional Techniques – Decision Criteria (Peter Welker, Senior Principal Consultant)
I was testing Oracle Goldengate on a non-clustered Oracle 18.104.22.168 database with ASM. With ASM, you need to have the grid infrastructure installed. The cluster ware for the single node install is called ‘oracle restart’.
The most convenient way to have Goldengate running at startup that I could find, was using the Oracle Grid Infrastructure Agents. These agents are not installed by default, you need to download these from the Oracle Technology Network. The download is with the grid infrastructure downloads in the database section.
The installation is very simple: unzip the xagpack_6.zip file, and run the xagsetup.sh script with the arguments ‘–install’ and ‘–directory’ arguments. The directory argument needs to point to a directory outside of the grid infrastructure home. My first choice would be to have it in the infrastructure home, but the install script does not allow that.
Do you know the difference between exp/imp and expdp/impdp when it comes to importing HCC compressed data in Exadata?
If not, then follow me through two examples. This is on 22.214.171.124/126.96.36.199.1 but applies to all database releases you can have on Exadata. The task at hand is to export a table (which happens to be non-partitioned and HCC compressed for query high) and import it into a different user’s schema. This is quite a common approach when migrating data from a non-Exadata system into an Exadata system. You could for example pre-create the DDL for the tables and implement HCC before even importing a single row. When importing the data, the partitions’ HCC attributes will be honoured and data will be inserted compressed. Or won’t it?
You can now download and have another look at the presentations used during the In-Memory…