Create table space command with no data path specification.

I use the following command to create the table space in oracle 11g
TestTbSpace create TABLESPACE DATAFILE 'dataFileName.dbf' SIZE 50 M REUSE AUTOEXTEND ON NEXT 1 M MAXSIZE 32767 M NOLOGGING"
But he is creating datafile dataFileName.dbf to disk at the following path
echo $ORACLE_HOME/dbs
I don't want to create datafile to this path and do not want to specify the path to the data file in the command 'create a table space.

Is there is no parameter, what I put it and especially the start command to create a dataFileName.dbf to this path

Published by: user8680179 on May 15, 2012 01:54

That's what I tried-

SQL> show parameter create_file

NAME                                 TYPE        VALUE
------------------------------------ ----------- ------------------------------
db_create_file_dest                  string
SQL>
SQL> alter system set db_create_file_dest='D:\oracle\oradata';

System altered.

SQL> show parameter create_file

NAME                                 TYPE        VALUE
------------------------------------ ----------- ------------------------------
db_create_file_dest                  string      D:\oracle\oradata

SQL> create tablespace sub datafile *'sub_01.dbf'* size 5M;

Tablespace created.

SQL> @filename
Enter value for tbs_name: sub
old   1: select file_id,file_name,(bytes/1024/1024)MB,AUTOEXTENSIBLE, MAXBYTES/1024/1024 "MAX size Mb" from dba_data_files where tablespace_name=UPPER('&tbs_name')
new   1: select file_id,file_name,(bytes/1024/1024)MB,AUTOEXTENSIBLE, MAXBYTES/1024/1024 "MAX size Mb" from dba_data_files where tablespace_name=UPPER('sub')

   FILE_ID FILE_NAME                                                                      MB AUT MAX size Mb
---------- ---------------------------------------------------------------------- ---------- --- -----------
        11 *D:\ORACLE\PRODUCT\11.2.0.2\DBHOME_1\DATABASE\SUB_01.DBF*                         5 NO            0

SQL> create tablespace note datafile size 5M;

Tablespace created.

SQL> @filename
Enter value for tbs_name: note
old   1: select file_id,file_name,(bytes/1024/1024)MB,AUTOEXTENSIBLE, MAXBYTES/1024/1024 "MAX size Mb" from dba_data_files where tablespace_name=UPPER('&tbs_name')
new   1: select file_id,file_name,(bytes/1024/1024)MB,AUTOEXTENSIBLE, MAXBYTES/1024/1024 "MAX size Mb" from dba_data_files where tablespace_name=UPPER('note')

   FILE_ID FILE_NAME                                                                      MB AUT MAX size Mb
---------- ---------------------------------------------------------------------- ---------- --- -----------
        12 *D:\ORACLE\ORADATA\MATSTDBY\DATAFILE\O1_MF_NOTE_7V6B0W0Z_.DBF*                    5 NO            0

Anand

Tags: Database

Similar Questions

  • Cannot create a new startup with my data disc

    Hi guys, here's the problem: when I try to create a new startup with my data disc, I can't boot from this drive.

    The startup progress bar takes forever and then I see a white circle with a line through it. ("Prohibited" / "Ghostbusters" sign.)

    In verbose mode, it hangs on "break pci: SDXC ' for a while, then it clears the screen and watch the Circle + line with 'still waiting on the root device' at the bottom left.

    I can't boot into single user mode. Quite simply, there is the same thing as verbose mode. Even in safe mode. So I have few options for debugging. I did a SMC reset, no change. Is there anything else I should try?

    Additional information:

    -I'm trying to start off a new SSD that I bought, currently in a USB 3.0 enclosure external. I know that all is well with the speaker because I started off of him several times in the past. I know that all is well with the reader, because I tried to do the same process with another drive and it doesn't work the same way.

    -J' cloned my current boot drive to the new drive with SuperDuper!, Carbon Copy Cloner and I also did a clean install of El Capitan using recovery mode. The new facility has worked well until I migrated my old data.

    - So there must be something in my data which are striated throughout this process. I don't know what it is maybe or how to track him down.

    Any thoughts are appreciated!

    Have you tried the PRAM reset? Restart while pressing control + command + P + R until you hear the boot chime at least twice.

  • creating table of partition with null values

    Hello

    I have a table with 20 GB of data. I changed this table in the table to partition with partition of the range
    the column with which I took for range (date), holding the values null

    can I have a partition for null values?

    Hello

    CREATE TABLE ICBS. KNET_STMT_EXTRACT_NEW_P
    (
    SRVC_TYPE VARCHAR2 (200 BYTE),
    FILLER_1 VARCHAR2 (200 BYTE),
    RTE_STAT VARCHAR2 (200 BYTE),
    TRAN_DATE VARCHAR2 (200 BYTE),
    TRAN_TIME VARCHAR2 (200 BYTE),
    POST_DATE VARCHAR2 (200 BYTE),
    DATE OF REC_INSERT_DT,
    DATE OF PROCESSED_DATE,
    KNET_RECON_RESSON VARCHAR2 (100 BYTE),
    NEW_TRAN_DATE VARCHAR2 (10 BYTE),
    REASON VARCHAR2 (200 BYTE)
    REASON_IND NUMBER (5),
    NODE_ID VARCHAR2 (8 BYTE),
    PROGRAM_ID VARCHAR2 (8 BYTE),
    USER_ID VARCHAR2 (8 BYTE),
    TIME_STAMP DATE
    )
    ICBS TABLESPACE
    PCTUSED 0
    PCTFREE 10
    INITRANS 1
    MAXTRANS 255
    STORAGE)
    64K INITIALS
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    PCTINCREASE 0
    DEFAULT USER_TABLES
    )
    LOGGING
    PARTITION BY RANGE (REC_INSERT_DT)
    (

    KNET_STMT_EXTRACT_NEW_P_DEC11 PARTITION VALUES LESS THAN (TO_DATE (' 2012-01-01 00:00:00 ',' SYYYY-MM-DD HH24:MI:SS ',' NLS_CALENDAR = GREGORIAN '))
    LOGGING
    NOCOMPRESS
    ICBS TABLESPACE
    PCTFREE 10
    INITRANS 1
    MAXTRANS 255
    STORAGE)
    INITIAL 1 M
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    DEFAULT USER_TABLES
    ),
    KNET_STMT_EXTRACT_NEW_P_JAN12 PARTITION VALUES LESS THAN (TO_DATE (' 2012-02-01 00:00:00 ',' SYYYY-MM-DD HH24:MI:SS ',' NLS_CALENDAR = GREGORIAN '))
    LOGGING
    NOCOMPRESS
    ICBS TABLESPACE
    PCTFREE 10
    INITRANS 1
    MAXTRANS 255
    STORAGE)
    INITIAL 1 M
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    DEFAULT USER_TABLES
    ),
    PARTITION KNET_STMT_EXTRACT_NEW_P_MVAL VALUES LESS THAN (MAXVALUE)
    LOGGING
    NOCOMPRESS
    TABLESPACE TBL_MAXVALUE
    PCTFREE 10
    INITRANS 1
    MAXTRANS 255
    STORAGE)
    INITIAL 1 M
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    DEFAULT USER_TABLES
    )
    NOCOMPRESS
    NOCACHE
    NOPARALLEL
    MONITORING
    ALLOW THE MOVEMENT OF THE LINE;

    Published by: hitgon on April 30, 2012 10:56

    Published by: hitgon on April 30, 2012 10:58

    Published by: hitgon on April 30, 2012 10:59

  • Create Table in Access with AutoIncrement field

    Hello

    I like to create a new table in an access database.

    The problem is to create an ID field with autoincrement. I can add "not null" and "primary key" for the field, but not the option "autoincrement".

    You have a solution to this problem?

    Thank you

    Marcus

    Access autoincrement

    It seems that if the Autoincrement does not work, you can try Counter as specification of column.

    /Y

  • There is a problem of storage with the data path [path to hard] store

    I have two sites connected by using a VPN connection over a WAN link. In each site, I have an esxi with several VMS host and operation of replication device of vSphere (v 5.1.1.0 Build 1079383) in each site. In each site, these is also a vCenter server with the device of RVS registered and logged with the other site. All the virtual machines are configured to replicate to the other site.

    In my initial configuration of the RV (first install) site remote had 2 vNIC used for the management and I figured it was my problem with replication (see kb2040302). So I have changed since my host configuration from the remote site to have that single vnic management (does not solve the problem). I then registered, removed and reinstalled the VR device without change to the problem.

    What is going on:

    1. I am able to configure replication between sites without a problem.

    2. first replication occurs between sites and reports a success.

    3. the remote VM replication to the main site continue to work properly and all is well for the days and continue to work.

    4. the main site of VM replication to the remote site fails on the second and all attempts to reproduce the error. There is a problem of storage with datastore path '[datastorename] '.

    ReplicationError.JPG

    I followed the advice of kb2040302 better as I can. But I must confess that I could use some of the more detailed instructions on how and what site should I check the logs. and how to check my configuration management vnic.

    I'm looking for someone provide additional troubleshooting steps with good detail.

    Thank you


    Resolution:

    The problem was caused by a security setting on the NFS Server data store. VRA could not delete the (hard).

    Specifically, my problem was with authorizations on the Synology NAS under the application of the "file Station.

  • How to create an interactive PDF with variable data

    We basically have a "mail merge" to our list of customers with an interactive PDF file (including videos, menus, etc. - not only form fill and web links) to create a single PDF that contains several pieces of mail... one for each customer... with each total mail being customized for this customer.  Customizations would include different greetings (Bob dear, dear Dana, etc), as well as the different graphs based on unique data to the client, photographs, etc.

    I saw that InDesign and Acrobat Professional can be used to create an interactive PDF (such as http://tv.adobe.com/watch/ask-the-adobe-ones/14-calling-rufus-about-interactive-pdf-making).  But I don't understand how I can insert data into a database, csv, excel etc. in the PDF file file so that each page, or each set of pages in the PDF file can be customized.

    Can someone tell me a tool to use for this?

    Thank you

    Bob Kendall

    I'm new on this and didn't know InDesign could do a data merge, thank you for pointing that out.

    However, we currently produce hundreds of thousands of personalized daily electronic documents, a PDF file or Web user interaction (no, this isn't spam or junk mail).  Our customers have asked us to focus on the generation of personalized interactive PDFs and I fear that the volume could overwhelm InDesign.  My other concern is that all of our work performed automatically (we run hundreds every day) and it is not possible to have a person to sit down, then run InDesign manually for each job.  Our standard operating system need a certain version of unix (AIX or linux), although we can perform tasks under windows if necessary.  It seems, as InDesign will fit in this environment?  If not, is there something else out there?

    Thank you for your time.

    -Bob

  • Laptop HP Stream - 11-d010wm: having to check the compatibility of a MicroSD card with my data flow specifications before buying the card

    A SanDisk Ultra 16 GB UHS-I/class 10 Micro SDHC memory is not compatible with a laptop HP of Stream - 11-d010wm (ENERGY STAR)?  I have a couple of Micro SD cards, but they are getting older and being not recognized when I put them in the card slot reader, so I guess that they are not compatible.  I'm looking to buy a new SD card, but not to run into the same problem, so want to check compatibility before you buy... any help would be appreciated.  Thank you.

    You can also consult the many Amazon comments and answered questions for the 11 waterways, as some say the real SD card they bought (although most seem to just mention 'micro SD' so I really think that everything you buy is going to work).  I can't explain why the 2 you have now are not unless they received a different kind of formatting.

    Ah wait however, this review has reported problems with the three different SanDisk class 10 cards, maybe you can search through the comments and find some that work.

  • quickly create with filter-date and initialize it with the current date

    Oracle BI 11 g

    Hello!

    I need to create guest of dashboard with filter-date and initialize it with the current date. How can I do?

    I tried to create the repository initialization block and add a variable. But I don't know what should I write to DataSource? I tried using the Current_Date, Now(), sysdate functions (for example, SELECT Now() FROM tbl_Calendar) but without results - when I pressed the button "Test"... "I have errors - something like 'Now() is unknown function' or 'incorrect syntax near keyword Current_Date.

    After that, I made to use presentation Variable in the command prompt, but also without success (())

    Please, help me.

    Use "Server Variable.

  • Line chart of several series with staggered data points

    Hello

    App of Hillary graphic sample includes an examples of charts and how to handle null values.

    https://Apex.Oracle.com/pls/Apex/f?p=36648:55

    I came across a scenario which goes a step further. If you have a series with a set of data points and another series that overlap, where a few data points are missing, for example:

    create table graph_lines (series number, thedate date, thevalue number);
    insert into graph_lines
    select 1 series
      ,add_months(trunc(sysdate,'yy'),rownum-1) thedate
      ,log(2,rownum) thevalue
    from dual
    connect by level <= 12
    union all
    select 2
      ,add_months(trunc(sysdate,'yy'),rownum-1)
      ,case when mod(rownum,3) =0 or rownum =2 then log(3,rownum) end
    from dual
    connect by level <= 12;
    

    I get the case where he's going to draw a line between the points that are next to each other, but the gaps it will display just the point

    graph.png

    Https://apex.oracle.com/pls/apex/f?p=73000:35 image

    I have no doubt a single this is the expected behavior, but I would like to know if anyone has any ideas on how to connect the dots.

    I thought about a snazzy model clause generates middle in my data points might work, but this seems heavy handed and a little scary.

    I thought that maybe there is a graphic solution?

    APEX 4.2.x

    Scott

    The idea of 'Null' removal was not all that possible. After testing it out, I found apex will be always generated empty points. Explicitly define the XML to have only the data points with values does nothing either.

    Because I thought it was a really interesting problem, I was wasting so much time on this :-). I was side on the reference guide and the user, modify the XML again and again and again.

    I hoped that, by defining a second category axis and set regional and datetime parameters, change of the shaft in one type datetime and specifying a main grid with every month, the table would be plot my data on this axis. Unfortunately, it seems as if this is ignored - a very frustrating experience. Anychart would just keep trace points with its own calculated intervals.

    Ultimately if I stumpled on a sort of difficulty, although I'm not sure how good it would be. There is a setting that you can set by using the custom XML part that ignores the missing data and the values of interpolation. It works well, but it always displays labels.

    https://Apex.Oracle.com/pls/Apex/f?p=11031:4 apex_demo/demo

    AnyChart | [AnyChart 6.x] How to connect mind lines...

    node - reference XML - AnyChart.Com Flash Chart Component

    With all these nodes, is it easy to miss what you need... It's a bit weird, but who "ignore the missing" points, you must set the value to "false"? :-)

  • Nested table and tables nested inside (CREATE TABLE)

    Hi all

    I have a problem of creation of a table with some types defined by the user at several levels (and therefore the nested tables).

    All first of all objects:
    type top_object is object (
        text_top varchar2(100),
        table_a a_object_tbl,
        table_b b_object_tbl
    );
    
    type a_object_tbl is table of a_object;
    type a_object is object (
      text_a varchar2(100)
    );
    
    type b_object_tbl is table of b_object;
    type b_object is object (
      text_b varchar2(100),
      table_c c_object_tbl,
      table_d d_object_tbl
    );
    
    type c_object_tbl is table of c_object;
    type c_object is object (
      text_c varchar2(100)
    );
    
    type d_object_tbl is table of d_object;
    type d_object is object (
      text_d varchar2(100)
    );
    Shortly: the SUPERIOR object contains tables A and B. B includes tables C and D.


    Well, now I want to create a table that contains the TOP level object.
    My attempt:
    create table multilevel_table (
      some_value varchar2(100),
      data top_object
    )
    NESTED TABLE top_object.table_a STORE AS data1,
    NESTED TABLE top_object.table_b STORE AS data2
    (NESTED TABLE table_c STORE AS data3,
    NESTED TABLE table_c STORE AS data4);
    Of course, this does not (otherwise I would not have asked you ;)). If I jump the 'second level', so just deleting table_b top_object and, therefore, the "inner nested table' it works.
    But on this scenario more complex I don't know the right way or the syntax.

    I would be grateful if I get any help or suggestion on your part!

    + 1 for recommendation of Hoek of simply not using TABLEs IMBRIQUEES for persistent structures.

    But:

    create table multilevel_table (
      some_value varchar2(100),
      data top_object
    )
    NESTED TABLE data.table_a STORE AS data1,
    NESTED TABLE data.table_b STORE AS data2
    (NESTED TABLE table_c STORE AS data3,
     NESTED TABLE table_d STORE AS data4) ;
    
  • ESXi with two data warehouses

    Hello

    I created an ESXi server with 2 data warehouses (2 logical drives). on the logic reader1 esxi is installed, and datastore1 is located. in this datastore1 all evnironments tests are recorded. on the second locial drive2 datastore2 is registered. There are all produvtive stored systems.

    When I added the second esxi data store it says that all existing data on the logical drive will be removed. No OK no problem right now.

    now logic disk1 ranout (datastore1) of space (50 GB hard drive). So I decided to replace it with a new hard drive with a lot more capacity. During the procedure to install esxi on the new drive hard I won't lose all data on logical disk1 and datastore1. but again not problem - test only environments (2 machines) it.

    But if I add my second drive to esxi - will I lose all the data on it too? or is there a workaround for just add to esxi without losing all my data?

    Thanks for help.

    Best regards

    bxtux007

    You will have to do so at the console - http://www.vm-help.com/esx/esx3i/ESXi_enable_SSH.php.

  • How to get the default path of the data of a table space file

    I used below the sql statement to create a table space:

    CREATE A TABLESPACE DATAFILE AAA BBB SIZE 10 M AUTOEXTEND OFF...

    Notice, I do not give a full path instead of a file name, and then ask questions relevant to tables

    get the full path of the data file, assume that it's path below:

    d:\oracle\...\BBB

    So I don't think that there is a default path that oracle can use it to create the data file. My question

    is how to get this default path before creating a table space.

    Thank you very much.

    Hello

    I did similar simulations for you, then it could be seen easily:

    SQL > Show parameter db_create_file_dest

    VALUE OF TYPE NAME

    ------------------------------------ ----------- ------------------------------

    db_create_file_dest chain

    SQL > SELECT NAME, VALUE OF V$ PARAMETER

    2 where lower (name) = "db_create_file_dest;

    NAME

    --------------------------------------------------------------------------------

    VALUE

    --------------------------------------------------------------------------------

    db_create_file_dest

    It has the value NULL. Then where my datafile is going?

    SQL > create tablespace TEST datafile 'test01.dbf' size 10 M;

    Created tablespace.

    SQL > select file_name

    2 of dba_data_files

    3 where nom_tablespace = 'TEST ';

    FILE_NAME

    --------------------------------------------------------------------------------

    /U01/app/Oracle/product/12.1.0/Db_1/DBS/Test01.dbf

    When you check the alert logs, it will not show you the location of the data file. It will tell you what exactly you ran.

    ..

    ..

    Completed: drop tablespace test whose content and data files

    create tablespace datafile 'test01.dbf' size 10 M TEST

    Completed: create tablespace TEST datafile 'test01.dbf' size 10 M

    ..

    ..

    You can not SQL user * more? Well, I would say that, when you run the CREATE TABLESPACE command, is where you run the command:

    SQL > select file_name

    2 of dba_data_files

    3 where nom_tablespace = 'TEST ';

    FILE_NAME

    --------------------------------------------------------------------------------

    /U01/app/Oracle/product/12.1.0/Db_1/DBS/Test01.dbf


    So, you should get the location of your data file. Alternatively, you can run the query in the parameter $ v so that you know your default data file location.

    I hope this helps.

    Thank you.

    Kind regards

    Gaetan

  • Create table with overlapping parallel dates in individual columns

    I try to combine data from two different tables into a single table.

    The data in table 1 contains locations of patients in a hospital where each record represents a single location. Patients can be transferred several times between the different beds resulting from multiple records for a single visit.

    The data in table 2 contains the operative activity of the patient to the Hospital where each record represents either the GOLD of the suspension of the recovery room. A patient may have multiple operations in a single visit.

    I would like to join/merge/mashup data in a single table that contains the data parallel to each other. In other words, dates of the appliance on one side of the table and the activity of GOLD on the other. The difficulty is that the two sets of overlapping of dates of arrival and departure. I wish that the final table to divide the originals in new records when the overlaps do not coincide.

    Example:

    Original in both events (one per table)

    > Unit event has - from 14:00 to 18:00

    > OR event B - from 16:00 to 17:00

    Results in 3 documents (in the final)

    > Event 1 - unit from 14:00 to 16:00, null dates GOLD

    > Event 2 - unit from 16:00 to 17:00 OR 16:00 to 17:00

    > Event 3 - unit from 17:00 to 18:00, null dates GOLD

    Of course overlap can be more complex than the example above and adding code to indicate the 'ghosts' transfers to as well.

    In the code below, the first visit of the GOLD occurs during the first mention of the unit.

    Jason

    Oracle 10g

    [code]

    create the table delme_Unit_dates

    (id varchar2 (20))

    , unit_rcd_id varchar2 (20)

    , Unit_desc varchar2 (20)

    Unit_in_code char (1)

    Date of Unit_in_dttm

    Date of Unit_out_dttm

    Unit_out_code char(1));

    create the table delme_or_dates

    (id varchar2 (20))

    , OR_rcd_id varchar2 (20)

    , OR_desc varchar2 (20)

    OR_in_code char (1)

    Date of OR_in_dttm

    Date of OR_out_dttm

    OR_out_code char(1));

    create the table delme_all_dates

    (id varchar2 (20))

    , Unit_OR_id varchar2 (40)

    , Unit_rcd_id varchar2 (20)

    , Unit_desc varchar2 (20)

    Unit_in_code char (1)

    Date of Unit_in_dttm

    Date of Unit_out_dttm

    Unit_out_code char (1)

    , OR_rcd_id varchar2 (20)

    , OR_Desc varchar2 (20)

    OR_in_code char (1)

    Date of OR_in_dttm

    Date of OR_out_dttm

    OR_out_code char (1));

    insert into delme_unit_dates values ('123456', 'U1111', 'Unit A', 'A', to_date('2013-04-29 5:02:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-09 1:06:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'B');

    insert into delme_unit_dates values ('123456', 'U1112', 'Unit A', 'B', to_date('2013-05-09 1:06:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-09 4:53:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'B');

    insert into delme_unit_dates values ('123456', 'U1113', 'Unit A', 'B', to_date('2013-05-09 4:53:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-10 10:52:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 't');

    insert into delme_unit_dates values ('123456', 'U1114', ' unity, 't', to_date('2013-05-10 10:52:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-11 11:30:00 AM', 'yyyy-mm-dd hh:mi:ss am'), 'B' ");

    insert into delme_unit_dates values ('123456', 'U1115', ' unity, ' B', to_date('2013-05-11 11:30:00 AM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-12 4:00:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'B');

    insert into delme_unit_dates values ('123456', 'U1116', ' unity, ' B ', to_date('2013-05-12 4:00:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-16 2:14:00 PM', 'yyyy-mm-dd hh:mi:ss am'),' t ');

    insert into delme_unit_dates values ('123456', 'U1117', 'Unit Z', ', to_date('2013-05-16 2:14:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-17 2:26:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'B');

    insert into delme_unit_dates values ('123456 ', 'U1118', 'Unit Z', 'B', to_date('2013-05-17 2:26:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-20 11:30:00 AM', 'yyyy-mm-dd hh:mi:ss am'),');

    insert into delme_or_dates values ('123456', 'OR2221', 'or 1', 'O', to_date('2013-05-09 7:35:00 AM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-09 10:56:00 AM', 'yyyy-mm-dd hh:mi:ss am'), 'R');

    insert into delme_or_dates values ('123456', 'OR2222', ' 5', 'R', to_date('2013-05-09 10:56:00 AM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-09 3:20:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'U');

    insert into delme_or_dates values ('123456', 'OR3331', 'or 2', 'O', to_date('2013-05-16 7:59:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-16 10:43:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'R');

    insert into delme_or_dates values ('123456', 'OR3332', ' 8', 'R', to_date('2013-05-16 10:43:00 PM', 'yyyy-mm-dd hh:mi:ss am'), to_date('2013-05-17 11:20:00 PM', 'yyyy-mm-dd hh:mi:ss am'), 'U');

    commit;

    -Is far from what we

    Select

    U.*

    , o.*

    Of

    delme_Unit_dates U

    delme_OR_dates O

    where

    U.ID = o.id

    and U.UNIT_IN_DTTM < = O.OR_IN_DTTM

    and U.UNIT_OUT_DTTM > = O.OR_IN_DTTM

    order of U.UNIT_IN_DTTM, O.OR_IN_DTTM

    ;

    [/ code]

    Post edited by: Jason_S (changed a single date ' 2013 - 05 - 16 15:20 ' to ' 2013 - 05 - 09 15:20 ')

    Hi, Jason.

    Jason_S wrote:

    I edited one of the dates in the original post.

    Also although the inpatient unit and OR events are contiguous for a given patient (no overlap and without gaps - after that data are cleaned).

    ...

    The sample data you posted a of gaps in the data of the GOLD.  It is correct that the solution below works or not there are gaps in the two tables.

    WITH got_dttm AS

    (

    SELECT unit_in_dttm AS DTMC

    OF delme_unit_dates

    UNION

    SELECT unit_out_dttm AS DTMC

    OF delme_unit_dates

    UNION

    SELECT or_in_dttm AS DTMC

    OF delme_or_dates

    UNION

    SELECT or_out_dttm AS DTMC

    OF delme_or_dates

    )

    all_periods AS

    (

    SELECT DTMC AS in_dttm

    , (DTMC) ahead OF (ORDER BY DTMC) AS out_dttm

    OF got_dttm

    )

    SELECT NVL (u.id, o.id) as id

    u.unit_rcd_id

    u.unit_desc

    u.unit_in_code

    p.in_dttm

    p.out_dttm

    o.or_rcd_id

    o.or_desc

    Of all_periods p

    LEFT OUTER JOIN delme_unit_dates u WE u.unit_in_dttm<=>

    AND u.unit_out_dttm > = p.out_dttm

    LEFT OUTER JOIN delme_or_dates o WE o.or_in_dttm<=>

    AND o.or_out_dttm > = p.out_dttm

    WHERE p.out_dttm IS NOT NULL

    ORDER BY p.in_dttm

    ;

    You can use the query above to CREATE or a CREATE TABLE... AS command.

    If you have as much data as you say, a table or materialized view would be maybe faster to use.

    You will notice that I do understand not all columns; I would like to know if you have a problem, including them.

    I don't know what id role plays in this problem.  It is difficult to say when all rows have the same value.

  • Create table with several Collections of data control

    Hello

    I have a requirement about to display data in a table. It shows simple, but certain columns in the table are within a collection, and the others are in another collection. I read this chapter Creation of ADF Databound Tables - 11 g Release 1 (11.1.1.7.0), and she explains to create the table with a collection.

    I have no knowledge on the model, I know just as control data, so what should I do? :

    -Ask for the template designer to redraw the control of data to include a collection with all the data?

    -Try to work with the actual data control? If so, how can I create the table several Collections of data control?

    Plese let me know if there is another solution. I work with JDeveloper 11.1.1.7.

    Thank you!

    AAPDL

    Hello

    You can have a managed bean that e... g exposes an ArrayList of objects. This can then binds to the value property of the table. If the object sets the information of table row combination that you get multiple iterators (collections), then the table will display the joined data.

    Frank

  • Create table with data in the column

    Create a new table, just want to know if there is a way to add a new column to the table with a value in all areas of this column when new rows are added to that this column will always be the same value

    As...

    name | address | zip | assets



    Active will always be Yes.
    I do a trigger?

    Use the default...

    create table (test)
    name varchar2 (20).
    address varchar2 (40),
    zip number (7).
    Active VARCHAR2 (3) DEFAULT NULL NOT 'yes');

Maybe you are looking for