Column appear in the Oracle data store
Hi guys,.My source and target are Oracle.
For the source, we have created the synonym of table views. Target is Oracle table. While I m try to overthrow the synonmys by selecting the synonmy option in the opposite selective tab of the model. The synonmys are getting reversed, but there is no column to view inside the data store.
Can someone help, how to reverse the synonmys Oracle in ODI which will serve as my source.
Thanks in advance.
There is a tab named 'Properties', it's just after the JDBC tab
Thank you
Fati
Tags: Business Intelligence
Similar Questions
-
Migrated virtual machine appears in the two data stores
Hi all
A bit of background on our installation first. servers ESX 3.5 vcenter 4.0 (just improved 2.5) + 2 + 2 data warehouses configured in 1 box of MSA.
Here is what happened. I've migrated a VM from A to B data store using "Migrate" in vSphere client on vCenter.
On the vSphere client, virtual machine displays now two data warehouses. In the browser data store, the vme even appears in the two data stores. Datastore shows 20 GB to 20 GB provisioned and 0 GB to 20 GB used. B displays 20 of 20 put into service and used 20 20 GB.
Migration seems to have succeeded. I migrated about 10 other vm and they reside only on a data store. All virtual machines are working properly. I tried to migrate the computer back to A virtual. He showed only a data store. But the problem appeared again when I migrated the vm to B. I tried to move the virtual machine back and forth between servers ESX, no difference.
Any ideas? The virtual machine is working well, and it is not necessarily a critical problem. But it's annoying, because I intend to reorganize the data store has and the fear it could cause a problem.
Thank you so much in advance!
anything mounted on the CD/DVD drive, as an ISO which can be on A Datastore?
-
Timestamp in the default data store is logical length of 13
I inserted a data store in the designer, the table lists fields that are TIMESTAMP (6) WITH ZONE SCHEDULE
The column of the logical data store default length is 13.
Execution fails with ORA-30088 because the ddl of the temporary table is TIMESTAMP (13) WITH ZONE SCHEDULE
I fixed the problem by changing the 05:47 on each column.
I want to know is how can I change the default to 6?
Or why is ODI not pick up it correctly?
Is there any solution for this?Hello
Topology of Goto---> physical architecture---> Oracle---> expand the types of data and change the TIMESTAMP WITH time ZONE SCHEDULE
See if the following is specified by report "Create table syntax" and "type of data accessible by writing syntax".
TIMESTAMP WITH TIME ZONE (%)If so, then change the same and remove (%).
that is, it will become TIMESTAMP WITH time ZONE SCHEDULESave and run your interface.
Thank you
Fati -
Filter on the Source data store
Hello
I am new to ODI. My source and target are the two Oracle. I have a table in the Source that has 10000 files. I Reversed engineer the source table and was able to view the data on the source data store. However, I want to filter the data on the source to send only rare recordings and not all 10000. I am writing a filter on the Data Source on a particular column store, but I still see all the records when I click Show data. Any suggestions?
Thank you
Arun
Edited by: user9525002 May 19, 2010 09:26
Edited by: user9525002 May 19, 2010 09:54Arun,
I don't think it's possible. You want to look at the filtered source data before loading to make sure that only correct data is loaded into the target.
A little more complicated would be to create a temporary interface (interface yellow) by selecting the Sunopsis Memory engine than the target temporary table.
And then right-click to display on this memory temporary table data. -
Not able to start agent cache for the requested data store
Hello
This is my first attempt in TimesTen. I am running TimesTen on the same host Linux (RHES 5.2) running Oracle 11 g R2. TimesTen version is:
TimesTen Release 11.2.1.4.0
Trying to create a simple cache.
The DSN entry section for ttdemo1 to. odbc.ini is as follows:
+ [ttdemo1] +.
Driver=/home/Oracle/TimesTen/TimesTen/lib/libtten.so
Data store = / work/oracle/TimesTen_store/ttdemo1
PermSize = 128
TempSize = 128
UID = hr
OracleId = MYDB
DatabaseCharacterSet = WE8MSWIN1252
ConnectionCharacterSet = WE8MSWIN1252
With the help of ttisql I connect
Command > Connect "dsn = ttdemo1; pwd = oracle; oraclepwd = oracle;
Successful login: DSN = ttdemo1; UID = hr; DataStore = / work/oracle/TimesTen_store/ttdemo1; DatabaseCharacterSet = WE8MSWIN1252; ConnectionCharacterSet=WE8MSWIN1252;DRIVER=/home/oracle/TimesTen/timesten/lib/libtten.so;OracleId=MYDB; PermSize = 128; TempSize = 128; TypeMode = 0; OracleNetServiceName = MYDB;
(Default AutoCommit = 1).
Command > call ttcacheuidpwdset ('ttsys', 'oracle');
Command > call ttcachestart;
* 10024: could not start agent cache for the requested data store. Could not initialize Handle.* Oracle environment
The command failed.
The following text appears in the tterrors.log:
15:41:21.82 Err: ORA: 9143: ora-9143 - 1252549744-xxagent03356: database: TTDEMO1 OCIEnvCreate failed. Return - 1 code
15:41:21.82 Err: 7140: oraagent says it failed to start: could not initialize manage Oracle environment.
15:41:22.36 Err: 7140: TT14004: failed to create the demon TimesTen: couldn't reproduce oraagent for "/ work/oracle/TimesTen_store/ttdemo1 ': has not been initialized Handl Oracle environment
What are the reasons that the demon cannot happen again to another agent? FYI, the environment variables are defined as:
ORA_NLS33=/U01/app/Oracle/product/11.2.0/Db_1/ocommon/NLS/Admin/data
ANT_HOME = / home/oracle/TimesTen/ttdemo1/3rdparty/ant
CLASSPATH=/home/oracle/TimesTen/ttdemo1/lib/ttjdbc5.jar:/home/oracle/TimesTen/ttdemo1/lib/orai18n.jar:/home/oracle/TimesTen/ttdemo1/lib/timestenjmsxla.jar:/home/oracle/TimesTen/ttdemo1/3rdparty/jms1.1/lib/jms.jar:.
Oracle@rhes5:/Home/Oracle/TimesTen/ttdemo1/info% echo $LD_LIBRARY_PATH
/home/oracle/TimesTen/ttdemo1/lib:/home/oracle/TimesTen/ttdemo1/ttoracle_home/instantclient_11_1:/u01/app/oracle/product/11.2.0/db_1/lib:/u01/app/oracle/product/11.2.0/db_1/network/lib:/lib:/usr/lib:/usr/ucblib:/usr/local/lib
See you soonI see no problem here. The ENOENTs are superfluous because it locates libtten here:
23302 open ("/ home/oracle/TimesTen/11.2.1.4.0/TimesTen/ttimdb1/lib/libtten.so", O_RDONLY) = 3
without doubt, it does the same thing trying to find the libttco.so?
23302 open ("/ home/oracle/TimesTen/11.2.1.4.0/TimesTen/ttimdb1/ttoracle_home/instantclient_11_1/libttco.so", O_RDONLY) =-1 ENOENT (no such file or directory)
Thank you for taking the trace. I really want to have a look at the complete file if you can send it to me?
-
Deployment uses the local data store
The deployment of VIO works without problem, but one of the virtual machines (VIO-DB-0) is now in a local data store. I never chose this data store in the installation process.
Is it possible to manually move the virtual machine? Is it possible to define the data store used for the management of virtual machines somewhere?
Concerning
Daniel
Hi Daniel,.
Take a look at VMware integrated OpenStack 1.0 Release Notes:
- Installer gives priority to local storage by default
When you set up the data for the database stores three virtual machines, VMware OpenStack Setup integrated automatically gives priority to a local storage to improve IO performance. For resilience, users might prefer a shared storage, but the installer does not clear how to change this setting. - Workaround: Before completing the installation process, the installer of VMware OpenStack integrated allows you to examine and change the configuration. You can use this opportunity to change the configuration of the data to the database store three virtual machines.
If you have already installed VIO, AFAIR you should be able to manually move between data warehouses as long as you do not change VM itself.
Note that there is rule anti-affinite, so you may not be able to move that VM in the same data store when no other VIO - DB resides.
Another note is that if you turn off VIO-DB-0, mysql service will not appear automatically. you have to turn on manually by running "service mysql start" on the node itself, or run 'vioconfig start' of the SGD server.
Best regards
Karol
- Installer gives priority to local storage by default
-
Unable to read / write in the NFS data store
Hi all
I'm having a problem with a NFS datastore in Vcenter. I have an NFS share on a Win2k3 server which I am able to mount. However... I can't write him even though the permissions appear to be correct. This server is connected to an array of training Vault EVA Storage with 2 TB of storage.
Looks like he's not more as when I try to mount ad content, it shows as 0.00 B ability... and I know that there are files in the NFS data store. Same Host, I can mount my other NFS datastore successfully since a 2008 Server, just do not know which cannot be configured correctly here.Help, please... been at this for 2 days banging my head on my desk!
Screenshots are attached. If there are log files that I could post that would help, please let me know and I'll join them as well.Thank you!Woot! It worked. I changed the local article, but then authorized security strategy anonymous access with the 2 GID and UID - 2. Once this is done, it is now properly show in Vcenter and allow me to read and write in the data store. Before that I got the value 0.0 even with security policy, change that did not work. So I hope that this issue helps someone in the future, and they do not have a hair pulling experience like I did.
WHAT A PAIN!
in any case, thank you all for your time.
-
ESXi 4.1 reinstall without loss of data in the system data store
Hello world
because of incorrect tape device configuration (see more low http://communities.vmware.com/thread/289847?tstart=0), I have to reinstall Esxi 4.1 in the same place where the system partition is also home to the Oracle data disk (I didn't have the choice because the system starts on the table of great raid with SSDS, there was no such thing which allow me to pre partition the RAID disk into several partitions and choose) only a small partition to install ESXi files), what happens if I reinstall ESXi with the method of "REPAIR"? will be large files within the system data store destroyed or intact? Help, please. Any suggestions are greatly appreciated.
When you do a normal installation of ESXi, the whole installation drive is wiped regardless of the existing partitions that exist. With a repair only install ESXi system partitions are left. Any data store must be perserved and after the repair, you can save any virtual machines that were welcomed on this data store. If the repair not proceed with the installation (for example, if the data store exists in the first 900 MB of storage), then you will get a warning in this sense and a new facility (wipe the entire disk) is performed.
So with a repair, you should be fine, but I want to take a backup without worrying.
Dave
VMware communities user moderator
Now available - vSphere Quick Start Guide
You have a system or a PCI with VMDirectPath? Submit your specifications to Officieux VMDirectPath HCL.
-
Display of the VM in the incorrect data store
Hello
I have 4 data warehouses and I've migrated a few VM data (DS) to another store, but when I discovered the 'data warehouses' in client VI still the virtual machine appears in the list on the old data store.
I have traveled the old DS and I can confirm no there are the virtual machine and the .vmx file points to the new location in the settings of the virtual machine, so why the virtual machine appears in the list of the old (source) DS?
Do you have something (for example ISO) attached to the virtual machine that may still reside on the old data store?
Review the vmx file (downloadable from the browser data store) and check that there is no reference to the old data store.
André
-
Hello experts,
I've created an interface when an xml file is reversed in the form of a source data store. Xml data are pumped into a target of oracle db. All this goes well.
I'm creating a scenario where I get an xml file from a ftp server on a daily basis (with agent). This new xml file has the same structure as that already used in the interface. The question is: How can I update the data in the xml data store?
I tried to replace the original xml file, but it does not work and cdc does not seem to apply here, I searched for quite a while now.
Thank you very much!
YvesHello
See if that helps
XML for the interface Oracle even insert County regardless of input XML fileThank you
Fati -
Good way to use the concurrent data store
Hello
I'm developing a multithreaded C++ application that uses the C++ of Berkeley DB Library.
In my case, I have several databases that I composed in a wooded area. It is important for me to use an environment because I need control over the cachesize parameter.
I don't need no guarantee of transaction and have for most of the readings, so I decided to use the "simultaneous database.
I first pre-fill all databases with a number of entries (configuration single-threaded phase) and then work on it at the same time (for most readings, but also insertions and deletions).
I tried all kinds of different configurations, but I can't work without specifying DB_THREAD as a flag of the environment.
I don't want to because then access all handles is synchronized sequentially according to the documentation:
"... Note that the activation of this indicator will serialize calls to DB using the handle between the threads. If
simultaneous scaling is important for your application, we recommend handles separate for each thread opening
(and do not specify this indicator), rather than share handles between threads. "
So I tried to open the environment with the following indicators:
DB_CREATE | DB_PRIVATE | DB_INIT_MPOOL | DB_INIT_CDB
All data in this environment handles are open only with the DB_CREATE flag.
So, since my understanding this same basic access handles need to be synchronized, I opened separate handles foreach database for each thread (opening the handles is still single-threaded).
In my first approach, I have only made use of the object of global environment. Which does not work and gives the following during operations error message:
DB_LOCK-> lock_put: Lock is no longer valid
So I thought, since the same handle global env is passed to all handles separate DB, it is perhaps a race condition critical on the handful of approx.
So in my next test, I opened also handles separate EPS in each thread (and db handles owned each).
That does not produce an error in db, but now it seems that each thread sees its own version of the databases (I call it stat early in the life of each thread and he sees all of the empty dbs).
What is the right way to use the concurrent data store? Each thread should really open his own set of db handles? What about the number of open handles env?
PS: Without specifying that the DB_PRIVATE flag seems to do the job, but for performance reasons, I want all operations to perform in the cache and do not specify product DB_PRIVATE average of several writes to the disk for my scenario.
Thanks a lot for your help.
CD (simultaneous database) allows a single editor with multiple drives, access to the db at a given point in time. The handle for the writer doesn't have to be shared with readers. If you share the DB handle then calls are synchronized, but if each thread has its own handle DB then this is not the case. Since you have an environment, DB_THREAD must be at the level of the environment. This will allow the sharing of the environment handle. This type of error "DB_LOCK-> lock_put: Lock is no longer valid" you can provide us your code so we can take a look. Also what BDB version are you using?
-
Need to create a structure for the target data store?
Hi Experts,
If I create a structure for the target data store, and then load the data from source to target works fine. If I make mistakes.
Is necessary to create a structure for target?
Please help me...
Thanks in advance.
A.Kavya.
I found the answer. No need to create the structure for target temporary data store. and we need to create the structure for the permanent target data store.
-
Move-VM error: VM must be managed by the same VI servers as the destination data store
I work with 2 ESXi servers and a single vCenter server with vSphere 6.0 Update 1 b; and PowerCLI 6.0 Release 2
I have an a datacenter, cluster, vCenter and a vDS. I need to pass the vCenter data store to another and get the error message:
"Move-VM VM you are moving should be handled by the same server of VI as the destination container and the data store.
Code looks like the following:
$vctObj = to connect-viserver $vctIPaddress
$esx1Obj = to connect-viserver $esx1IPaddress
$esx2Obj = to connect-viserver $esx2IPaddress
$vmObj = get-vm-name $vmname - Server $vctObj
$destDSobj = get-datastore-name $destDSname
Move-vm - vm $vmObj Datastore - $destDSobj - 'End' DiskStorageFormat
I can migrate successfully from the virtual machine to vCenter data store through the VI Client, but PowerCLI left me speechless.
Maureen
Try to use the Server parameter on the Get-Datastore line
-
remove the Cluster data store data store
I have an infrastructure with vCenter and ESXi 4 5.5 I have a data cluster store in SAN with 8 Mon, I need to remove 3 Lun (to be used for other purposes) what is the appropriate procedure
to remove the Lun (end then to destroy)? Thank you
Do you want the LUN to use for purpose of non-vSphere? If so, you can just storage vMotion virtual machines since associated LUN data warehouses that you want to decommission (or simply putting the data store in maintenance mode, in this way, that the virtual machines will automatically be migrated). Cleaning after the data store, move the data store from the cluster data store, and then delete the VMware environment data store as described here: best practices: how to properly remove a unit number logic of a host ESX - VMware vSphere Blog - VMware Blogs
-
Extremely high latency during the migration of the local data store to data store shared.
Hi guys I hope you help me. Sorry for my English btw, I'm not native.
Let's start!
I have:
1 vCenter
1 host
1 Distributed Switch (with a pg for managing network/IPstorage the esxi)
1 standard switch (empty)
1 FreeNAS to provide iSCSI LUNS
1 Microsoft to provide iSCSI LUNSWhen I try to migrate virtual machines between the warehouses of shared data or data shared at local store, all right. The problem has become when I try to migrate virtual machines from the local to the shared data store. All data stores down (all paths down) and back up and I get this error:
"Error caused by the /vmfs/volumes/volumenID/VMDirectory/Disk.vmdk file.
When I try to migrate virtual machines from the local to the FreeNAS iSCSI data store it fails immediately.
When I do the same local Microsoft iSCSI datastore that it takes a time loooooong to migrate the virtual computer, give it to me even all the paths downwards and uplinks down error but not lacking in migration.I'll give you some screenshots to see the errors.
Thank you very much!
EDIT: I have extremely high latency notice when I try to migrate from the premises of warehouses of shared data. average 2000ms with peaks of 50,000 (see my response below for more information)
Finally, I found the solution! The problem is that I have been using E1000E vmnic instead of vmxnet3. I have configured an adapter vmxnet3 and boom! 20 ms in all migration!
Thank you to all for help, especially Nick_Andreev !
Maybe you are looking for
-
Time Machine impossible to restore backups
Hello I had to restore my entire machine (because I am stupid) and when I went to access my Visual readers of Parallels, missing 2. It seems they were always absent (I went back I can. I'm not saying TM to ignore readers, and I see no reason why they
-
Satellite Pro L40 PSL43E - need Bios for Win XP
Need Bios Update for Windows XP to my laptop Satellite Pro L40 PSL43E Ty
-
Hello! I'm new to libraries of LV and I am compiling my application for the first time with several new lvlib. The executable is apparently built without problem, but when I try to run it on the same computer, I get various error messages (4) on the
-
I'm going on a plane later in the day and I need to use my laptop to type some work. My laptop Toshiba, Windows 7, I can't seem to understand how to activate flight mode can someone help me? See you soon
-
Smartphone blackBerry, Reboot, Reboot, Reboot
My "BOLD" continues to reboot sometimes 2-3 times a day. It goes right down and that returns to the top. I also get calls like crazy. Sometimes, I'll be on the phone and it restarts automatically? Any ideas or anyone with similar problems.