question of privilege apex data load

Hi all

I use apex to load a csv file into oracle11gr2

However, the table is created under the schema user apex and the apex user cannot see the table.

This isn't the same behavior as in apex.oracle.com or they use a trigger to create synonym and synonym for this table give?

Y at - it a standard package that I can run for this problem?

-Thank you very much for the help

still need help I mark as answer by mistake!

Published by: xwo0owx on March 29, 2011 15:19

Yes, we can create a means for a trigger like this

Create or replace trigger test before insertion on emp
Start
dbms_output.put_line ('Hi');
end;
Create the trigger in the scott schema

and connect to the apps schema and create a synonym like this

Create public synonym test for scott.test;

or

create synonym test1 for scott.test;

And also see this example

I created a trriger create, in which i'm giving execute immediate command for a synonym for creation and use
dbms_job. Submit.
It gives no error, but its not to create any synonym on the table...
Code:

create or replace trigger bdesai. DR_CRA
After you CREATE or DROP on bdesai.schema
declare
l_str VARCHAR2 (255);
l_str1 VARCHAR2 (255);
number of l_job;
Start
CASE ORA_SYSEVENT
WHEN "CREATE" CAN
If (ora_dict_obj_type = 'TABLE') then
l_str: = ' immediate "grant select on bdesai."
|| ora_dict_obj_name: ' to eqhub_developer ';
l_str1: = ' immediate ' create or replace synonym.
eqhub_developer' | '.'|| ora_dict_obj_name |' for
bdesai' | '.'|| ora_dict_obj_name |' » ; " ;
DBMS_OUTPUT. Put_line (replace (l_str,'"'," ' "));
DBMS_OUTPUT. Put_line (replace (l_str1,'"'," ' "));
dbms_job. Submit (l_job, replace(l_str,'"','''));
dbms_job. Submit (l_job, replace(l_str1,'"','''));

END IF;
WHEN CAN 'DROP'
If (ora_dict_obj_type = 'TABLE') then
l_str: = ' immediate execution "revoke select on bdesai."
|| ora_dict_obj_name: ' to eqhub_developer ';
l_str1: = ' immediate execution "drop synonym".
eqhub_developer' | '.'|| ora_dict_obj_name |' » ; " ;
DBMS_OUTPUT. Put_line (replace (l_str,'"'," ' "));
DBMS_OUTPUT. Put_line (replace (l_str1,'"'," ' "));
dbms_job. Submit (l_job, replace(l_str,'"','''));
dbms_job. Submit (l_job, replace(l_str1,'"','''));
END IF;
END CASE;

end;
/

Source: http://www.allinterview.com/showanswers/72132.html

Hope this helps you :)

Thank you

Barry K.

Tags: Database

Similar Questions

  • Apex data load configuration missing table when importing to the new workspace

    Hi everyone knows this show before and have a work around.

    I export / import my request for a different workspace, and everything works fine except for 1 loading tables data.

    In my application, I use tables of data loading, and it seems that the latter do not properly Setup for the table object to load data. This causes my application to fail when the user tries to download a text file.
    Does anyone have a work around next to recreate the table of data load object?

    Breadcrumb: Components shared-> load data tables-> add modify data load table

    The app before exporting displays Workspace: OOS
    Single column 1 - CURRENCY_ID (number)
    Single column 2 - month (Date)

    When I import the app in the new workspace (OOS_UAT) data type is absent.
    Single column 1 - CURRENCY_ID
    Single column 2 - MONTH

    When I import the same workspace app: OOS I do not know this problem

    Version of the apex: Application Express 4.1.1.00.23

    Hi all

    If you run 4.1.1 it was a bug 13780604 (DATA DOWNLOAD WIZARD FAILED if EXPORTS of OTHER workspace) and have been fixed. You can download the fix for 13780604 (support.us.oracle.com) and the associated 4.1.1

    Kind regards
    Patrick

  • limiting number of column in the 4.1 apex data loader?

    Hi all!

    Is there a limit on the number of column in the APEX 4.1 data loading page?

    My DB object has 59 columns and they are all available in single colum drop of my load sample data table definition boxes.
    On page two of the wizard created database load pages "data/table mapping" columns only 45 is indicated. These columns are correctly inserted in my table. The last 14 columns are ignored.

    If anyone knows if there is a limitation and can it be extended?

    Thanks for any reply and cordially
    Kai

    No, I don't have a solution for this.

    Split the file within columns each, with the primary key that is repeated and then sting tables up to post upload could be easier than to use other roads.

    And then there's always the good old SQL Loader and external Tables. But the integration of these into the Apex is not easy because Apex runs on the server and the file is usually the disk of HARD local to the client.

    Kind regards

  • Schema name is not displayed in the data loading

    Hi all

    I'm trying to load a CSV file using oracle apex data loading option. The options are using a new upload (.csv) file and table. In the data load page, the schema name is not list my current schema because of which I could not not to download the CSV file.
    Can someone please help with that?


    I use apex oracle 4.1.1

    Concerning
    Rajendrakumar.P

    Raj,

    If it works on apex.oracle.com (4.2) and not in your case (4.1.1), my suspicion is that this is a bug that has been fixed in 4.2 APEX. Apart from upgrading your version 4.2 of the APEX, I'm not sure that there is not really a viable alternative.

    Thank you

    -Scott-

    http://spendolini.blogspot.com
    http://www.enkitec.com

  • Newbie sorry data-load question and datafile / viral measure

    Hi guys

    Sorry disturbing you - but I did a lot of reading and am still confused.

    I was asked to create a new tablespace:

    create tablespace xyz datafile 'oradata/corpdata/xyz.dbf' size 2048M extent management local size unique 1023M;

    alter tablespace xyz add datafile ' / oradata/corpdata/xyz.dbf' size 2048M;

    Despite being worried not given information about the data to load or why the tablespace must be sized that way - I was told to just 'do it '.

    Someone tried to load data - and there was a message in the alerts log.

    ORA-1652: unable to extend temp by 65472 segment in tablespace xyz

    We do not use autoextend on data files even if the person loading the data would be so (they are new on the environment).

    The database is on a cold backup nightly routine - we are in a rock anvil - we have no space on the server - to make RMAN and only 10 G left on the Strip for (Veritas) backup routine and thus control space with no autoextend management.

    As far as I know of the above error message is that the storage space is not large enough to hold the load data - but I was told by the person who imports the data they have it correctly dimensioned and it something I did when the database create order (although I have cut and pasted from their instructions - and I adapted to our environment - Windows 2003 SP2 but 32 bits).

    The person called to say I had messed up their data loading and was about to make me their manager for failing to do my job - and they did and my line manager said that I failed to correctly create the tablespace.

    When this person was asked to create the tablespace I asked why they thought that extensions should be 1023M and said it was a large data load that must be inserted to a certain extent.

    That sounds good... but I'm confused.

    1023M is very much - this means that you have only four extents in the tablespace until it reaches capacity.

    It is a load - is GIS data - I have not participated in the previous data loads GIS - other than monitor and change of tablespaces to support - and previous people have size it right - and I've never had no return. Guess I'm a bit lazy - just did as they asked.

    However, they never used 128K as a size measure never 1023M.

    Can I ask is 1023 M normal for large data loads - or I'm just the question - it seems excessive unless you really just a table and an index of 1023M?

    Thanks for any idea or other research.

    Assuming a block size of 8 KB, 65472 would be 511 MB. However, as it is a GIS database, my guess is that the database block size itself has been set to 16K, then 65472 is 1023MB.

    What load data is done? Oracle Export dump? Which includes a CREATE INDEX statement?
    Export-Import is a CREATE TABLE and INSERT so that you would get an ORA-1652 on it. So you get ORA-1652 if the array is created.
    However, you will get an ORA-1652 on an INDEX to CREATE the target segment (ie the Index) for this operation is initially created as a 'temporary' segment until the Index build is complete when it switches to be a 'temporary' to be a segment of "index".

    Also, if parallelism is used, each parallel operation would attempt to assign degrees of 1023 MB. Therefore, even if the final index target should have been only, say 512 MB, a CREATE INDEX with a DEGREE of 4 would begin with 4 extensions of 1023 MB each and would not decrease to less than that!

    A measure of 1023 MB size is, in my opinion, very bad. My guess is that they came up with an estimate of the size of the table and thought that the table should be inserted in 1 measure and, therefore, specified 1023 MB in the script that is provided to you. And it is wrong.

    Same Oracle AUTOALLOCATE goes only up to 64 MB extended when a Segment reached the mark of 1 GB.

  • Data loading Wizzard

    On Oracle Apex, is there a feasibility study to change the default feature

    1. can we convert the load data Wizard just insert to insert / update functionality based on the table of the source?

    2. possibility of Validation - Count of Records < target table is true, then the user should get a choice to continue with insert / cancel the data loading process.

    I use APEX 5.0

    Need it please advice on this 2 points

    Hi Sudhir,

    I'll answer your questions below:

    (1) Yes, loading data can be inserted/updated updated

    It's the default behavior, if you choose the right-hand columns in order to detect duplicate records, you will be able to see the records that are new and those who are up to date.

    (2) it will be a little tricky, but you can get by using the underlying collection. Loading data uses several collections to perform the operations, and on the first step, load us all the records of the user in the collection "CLOB_CONTENT". by checking this against the number of records in the underlying table, you can easily add a new validation before moving on to step 1 - step 2.

    Kind regards

    Patrick

  • Data loading convert timestamp

    I use APEX 5.0 and download data wizard allows to transfer data from excel.

    I define Update_Date as timestamp (2) and the data in Excel is 2015/06/30 12:21:57

    NLS_TIMESTAMP_FORMAT is HH12:MI:SSXFF AM JJ/MM/RR

    I created the rule of transformation for update_date in the body of the function, plsql as below

    declare

    l_input_from_csv varchar2 (25): = to_char(:update_date, 'MM/DD/YYYY HH12:MI:SS AM');

    l_date timestamp;

    Start

    l_date: = to_timestamp (l_input_from_csv, ' MM/DD/YYYY HH12:MI: SS AM' ");

    Return l_date;

    end;

    I keep having error of transformation rule.  If I do create a transformation rule, it will ask for invalid month.

    Please help me to fix this.

    Thank you very much in advance!

    Hi DorothySG,

    DorothySG wrote:

    Please test on demand delivery data loading

    I stick a few examples to the Instruction of Page, you can use copy and paste function, it will give the same result

    Please change a coma separator ',' on the other will be message "do not load".

    Check your Application 21919. I changed the rule of transformation:

    Data is loading properly.

    Kind regards

    Kiran

  • problems with the JSON data loading

    Hello

    I have follow-up Simon Widjaja (EDGEDOCKS) YouTube lesson for the JSON data loading external. But I am not able to connect at least the console database.

    I get this error: "error avascript in the handler! Type of event = element.

    Content.JSON is located in the folder. Data there are very simple:

    [

    {

    "title": "TITLE 1",

    'description': "DESCRIPTION 1"

    },

    {

    "title": "TITLE 2",

    'description': "DESCRIPTION 2"

    }

    ]

    And here's the code in edgeActions.js:

    (function ($, edge, compId) {})

    Composition of var = Edge.Composition, symbol = Edge.Symbol; alias for classes of edge commonly used

    Edge symbol: "internship."

    (function (symbolName) {}

    Symbol.bindElementAction (compId, NomSymbole, 'document', 'compositionReady', function (sym, e) {})

    external json data loading

    $.ajax({)

    type: 'GET ',.

    cache: false,

    URL: "content.json",

    data type: 'json ',.

    success: function (data) {console.log ("data:", data);},

    error: function() {console.log ("something went wrong") ;}}

    });

    });

    End of binding edge

    (}) ('step');

    End of edge symbol: "internship."

    }) (window.jQuery |) AdobeEdge. ($, AdobeEdge, "EDGE-11125477");

    I tried $getJSON also as mentioned in the youtube video.

    Please note: I do not understand 'something was wrong' also connected.

    I use the free trial version. It is a limitation in the free trial version?

    Well, same question as here: loading external data using ajax

    Cannot run the jQuery file is missing, then $. ajax() or $. getJSON().

    You must add the jQuery file as shown below:

    See: http://jquery.com/download/

    Note: Without loading the jQuery file, you can use these functions: API JavaScript Adobe Edge animate CC

  • Event scripts FDM shot twice during data loads

    Here's an interesting question. I added the following three scripts to different event (one at a time, ensuring that one of them is both), clear data before loading to Essbase:


    Script event content:
    ' Declare local variables
    Dim ObjShell
    Dim strCMD
    «Call MaxL script to perform data clear the calculation.»
    Set objShell = CreateObject ("WScript.Shell")
    strCMD = "D:\Oracle\Middleware\EPMSystem11R1\products\Essbase\EssbaseClient\bin\startMAXL.cmd D:\Test.mxl"
    API. DataWindow.Utilities.mShellAndWait strCMD, 0


    MaxL Script:
    Login * identified by * on *;
    run the calculation ' FIX("Member1","Member2") CLEARDATA "Member3"; ENDFIX' on *. *** ;
    "exit";




    However, it seems that clear is performed twice, both before and after the data has been loaded to Essbase. This has been verified at every step, checking the newspaper of Essbase applications:

    No script event:
    -No Essbase data don't clear in the application log

    Above to add the script to the event "BefExportToDat":
    -The script is executed once when you click Export in the customer Web FDM (before the "target load" modal popup is displayed). Entries are visible in the log of Essbase applications.
    -Script is then run a second time when you click the OK button in the modal pop-up "target Load System". Entries are visible in the log of Essbase applications.

    Above to add the script to the event "AftExportToDat":
    -The script is executed once when you click Export in the customer Web FDM (before the "target load" modal popup is displayed). Entries are visible in the log of Essbase applications.
    -Script is then run a second time when you click the OK button in the modal pop-up "target Load System". Entries are visible in the log of Essbase applications.

    Above to add the script to the event "BefLoad":
    -Script only runs that after you click Export in the FDM Web Client (before 'target system load' modal popup is displayed).
    -Script is run AFTER loading to Essbase data when the OK button is clicked in the modal popup "load the target system". Entries are visible in the log of Essbase applications.

    Some notes on the above:
    1. "BefExportToDat" and "AftExportToDat" are both performed twice, before and after the modal popup "target Load System". :-(
    2. "befLoad" is executed WHEN the data is loaded to Essbase. :-( :-(

    Someone please any idea how we could run a clear Essbase database before the data is loaded, and not after we have charged for up-to-date data? And maybe about why event scripts above seem to be fired twice? It doesn't seem to be any logic to this!


    BefExportToDat - entered in the journal Application Essbase:
    + [Sea 16 May 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.
    +...+

    + [Sea 16 May 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info (1003037) +]
    Updated load cells [98] data

    + [Sea 16 May 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info (1003024) +]
    Data load time: seconds [0.52]
    +...+

    + [Sea 16 May 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.


    AftExportToDat - entered in the journal Application Essbase:
    + [Sea 16 May 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.
    +...+

    + [Sea 16 May 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info (1003037) +]
    Updated load cells [98] data

    + [Sea 16 May 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info (1003024) +]
    Data load time: seconds [0.52]
    +...+

    + [Sea 16 May 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.


    BefLoad - entered in the journal Application Essbase:
    + [Sea 16 May 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.
    +...+

    + [Sea 16 May 16:23:44 2012]Local/Monthly/Monthly/admin@Native Directory/140095929911040/Info (1003037) +]
    Updated load cells [98] data

    + [Sea 16 May 16:23:44 2012]Local/Monthly/Monthly/admin@Native Directory/140095929911040/Info (1003024) +]
    Data load time: seconds [0.52]
    +...+

    + [Sea 16 May 16:23:45 2012]Local/Monthly/Monthly/admin@Native 140095860504320/Directory/Info (1013091) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:23:45 2012]Local/Monthly/Monthly/admin@Native 140095860504320/Directory/Info (1013162) +]
    + Received order [calculate] user [directory admin@Native] +.

    + [Sea 16 May 16:23:45 2012]Local/Monthly/Monthly/admin@Native 140095860504320/Directory/Info (1012555) +]
    + Erasure of the data in the partition [Member3] with fixed member [Period (Member1); Scenario (Member2)] +.

    James, the scripts export and the Load event will fire four times, once for each type of file: the. DAT file (main TB file),-A.DAT (log file),-B.DAT and - c.DAT.

    To work around this problem, then only run during the loading of the main TB file, add the following or something similar at the beginning of your event scripts. This assumes that strFile is in the list of parameters to the subroutine:

    Select Case LCase(Right(strFile,6))
         Case "-a.dat", "-b.dat", "-c.dat" Exit Sub
    End Select
    
  • Question of privileges...

    Hi all..


    Please help me with this question of privileges.

    Let's say that ""(schéma/utilisateur) User1 has a package "package1" that updates and inserts data into 'user1 tables' "»

    Now, if I "User2 (schema/user)" who has run privileged on user1.package.User2 has no any reply on the tables of user1.

    User2 can still run the package user1? (user2 has execute privilege on the package of user1 but doen't have)
    access the tables of user1)

    Thanks in advance...

    Hello

    Marella Phani wrote:
    Hi all..

    Please help me with this question of privileges.

    Let's say that ""(schéma/utilisateur) User1 has a package "package1" that updates and inserts data into 'user1 tables' "»

    Whenever you have any questions, please post whatever one needs to re - create the rpoblem and test their ideas. In this case, which includes a package (spec and body) and all tables that package references. Simplify as much as possible... You can probably do it with 20 lines of code.

    Now, if I "User2 (schema/user)" who has run privileged on user1.package.User2 has no any reply on the tables of user1.

    User2 can still run the package user1? (user2 has execute privilege on the package of user1 but doen't have)
    access the tables of user1)

    Yes, but don't take my word for it. Try it and see.
    This assumes that package1 is created with the default AUTHID DEFINE.

  • Migration of the APEX data to the local instance

    Hello

    I want my data in all the tables in my schema exported from the apex.oracle.com app and imported to another local instance, on a regular basis. Utility-> data loading/unloading feature Apex provide this functionality via XMLs but the process is manual and needs to be done table-wise and not schema-wise. I'm not sure of the CREATE DATABASE LINK function in this case, can it be useful? No other work around? Help, please.

    Thank you
    Mary

    Hello

    Yes, he is always there - I just tried. Are you clicking on the link or copy and paste it in the address bar? Sometimes, if you do this, you get leader / characters.

    Andy

  • Forcing errors when loading essbase nonleaf data loading

    Hi all

    I was wondering if anyone had experience forcing data load errors when loadrules is trying to push the nonleaf members data in an essbase cube.

    I obviously ETL level to prevent members to achieve State of charge which are no sheet of error handling, but I would however like the management of the additional errors (so not only fix the errors).

    ID much prefer errors to be rejected and shown in the log, rather than being crushed by aggregation in the background

    Have you tried to create a security filter for the user used by the load that allows only write at level 0 and not greater access?

  • Move rejected records to a table during a data load

    Hi all

    When I run my interfaces sometimes I get errors caused by "invalid records. I mean, some contents of field is not valid.

    Then I would move these invalid to another table records, while the data loading lights in order not to interrupt the loading of data.

    How can I do? There are examples that I could follow?

    Thanks in advance

    concerning

    Hi Alvaro,

    Here you can find different ways to achieve this goal and choose according to your requirement:

    https://community.Oracle.com/thread/3764279?SR=Inbox

  • The data load has run in Odi but still some interfaces are running in the operator tab

    Hi Experts,

    I'm working on the customization of the olive TREE, we run the incremental load every day. Data loading is completed successfully, but the operator status icon tab showing some interfaces running.

    Could you please, what is the reason behind still running the interfaces tab of the operator. Thanks to Advance.your valuable suggestion is very useful.

    Kind regards

    REDA

    What we called stale session and can be removed with the restart of the agent.

    You can also manually clean session expired operator.

  • Ignore the ASO - zero data loads and missing values

    Hello

    There is an option that ignores the zero values & the missing values in the dialog box when loading data in cube ASO interactively via EAS.

    Y at - it an option to specify the same in the MAXL Import data command? I couldn't find a technical reference.

    I have 12 months in the columns in the data flow. At least 1/4 of my data is zeros. Ignoring zeros keeps the size of the cube small and faster.

    We are on 11.1.2.2.

    Appreciate your thoughts.

    Thank you

    Ethan.

    The thing is that it's hidden in the command Alter Database (Aggregate Storage) , when you create the data loading buffer.  If you are not sure what a buffer for loading data, see loading data using pads.

Maybe you are looking for

  • Where are the pictures?

    When my iPhone 5s (iOS 9.3.2) is connected to my Windows 7 system and I use Windows Explorer to examine the contents of the iPhone, there is now only a subdirectory in the folder DCIM, 100APPLE, where there were once many of these subdirectories. In

  • The Magic Trackpad 2 will not pair correctly with a Mac Mini

    I just bought a 2014 2.8 GHz Intel Core i5, Mac Mini and the Magic Trackpad 2 but I can't match the trackpad. When I connect the trackpad by using the usb cable, the functions of a mouse, but the bluetooth does not. I caught a third party mouse, who

  • 1000 HP no. 11 mini laptop model: fatal system error mini computer laptop hp 1000

    I need the error code fatal system please

  • Create a LabVIEW step dynamically

    Hello I am quite new to Teststand. I am creating a step in LabVIEW dynamically using ActiveX in Teststand. What I am trying to specify the LabVIEW module using the TestStand API OR with the class 'Step' and the 'SpecifyModule' method It makes sense,

  • Full scale PXI - 6254 DAQmx Analog Input

    Hello I use PXI - 6254 Board to read the analog inputs. Configured channels using DAQmx create Channel.vi with sub parameters. In the configuration: CSR Min: 0 Max: 10 Units: Volt I read the channel using DAQmx Read U16 2D with the sample of 1. I exp