Update of field Notes from a csv file

Hello

I want to update the Notes field for all of the virtual machines in my environment from a .csv file.  To do this, I did and the export of my environment in a file .csv using this command:

Get - VM | Select Name, Notes | Export-Csv-path "c:\output\notes.csv" - NoTypeInformation

I now have a list of all virtual machines and their existing Notes.  I manually changed the notes fields keeping the existing descriptions and adding, for all the virtual machines that had no notes.  I would now like to merge the changes into vCenter and to crush any descriptions in vCenter but my code is any error on:

Import-Csv "c:\output\notes.csv" | % {Set - VM $_.} VMName-note $_. {Note - confirm: $false}

Any ideas?

Untitled.jpg

Well, the field of your CSV is 'Name', not "VMName".

So, try:

Import-Csv "c:\output\notes.csv" | % {Get - VM $_.} Name | Set-VM-note $_. {Note - confirm: $false}

Tags: VMware

Similar Questions

  • Configure vswitches from a .csv file - problem

    I have a script that works very well for the installation of my virtual past by using the updatevirtualswitch method. (Thanks to LucD, see here: ) http://communities.VMware.com/message/1556669#1556669 )

    I now want to go further and to feed my script with variables from a .csv file.

    It works fine for all, with the exception of the definition of vmnic. Description of the problem:

    Excerpt from my .csv file:

    lannics; dmznic1; dmznic2; dmznic3; storagenics; vmotionnics;

    @("vmnic0"); @("vmnic1", "vmnic2");

    Then import the settings from the csv file, example:

    $lannics = $parameterfile.dmznic1

    Now, if I check what's in $dmznic1, I have the good: @("vmnic1", "vmnic2")

    But it seems to be a string, not a true table. Therefore, I can not pass it to my updatevirtualswitch function:

    function standardvswitch {}

    Param ($esx, $vs, [string []] $dmznic1)

    ....

    $ns. UpdateVirtualSwitch ($vs, $vsSpec)

    }

    So the question is: how could I get my .csv file information, so that it can be used for a definition of vmnic compatible with the UpdateVirtualSwitch method?

    Thanks for your help or ideas!

    I've done a few tests more and saw that my previous solution does not work. "But one that follows). I'll try to explain how it works. In the .csv file, a semicolon is used as a field separator. This means that you can use a comma in a field to separate the members of the group. The Import-CSV cmdlet reads the .csv file and - separator "," parameter instructs the cmdlet that a semicolon is the field delimiter. The output of the Import-CSV cmdlet is piped in a ForEach-Object cmdlet. In the scriptblock ForEach-Object according to the value of the dmznic1 property string is shared by the comma, so each string before, between and after the comma will become a member of separate table. This table is assigned to the dmznic1 property. Loop loop through all the members of the Group and displays them on separate lines. So you can see that it is really a picture.

    Import-CSV -Path LanNics.CSV -Delimiter ";" | `
    ForEach-Object {
      $_.dmznic1 = ($_.dmznic1).split(",")
      for ($i=0;$i -lt $_.dmznic1.length; $i++) {
        Write-Output $_.dmznic1[$i]
      }
    }
    

    See the attached screenshot for output.

    I think that this solution is more beautiful that create a different column for each vmnic because in my solution you don't have to know in advance how many cards you have.

    Post edited by: RvdNieuwendijk

  • Unable to display data from a csv file data store

    Hi all

    I'm using ODI 11 g. I'm trying to import metadata from a csv file. To do this, I have created physical and logical diagrams corresponding. Context is global.

    Then, I created a model and a data store. Now, after reverse engineering data store, I got the file headers and I changed the data type of columns to my requirement and then tried to view the data in the data store. I am not getting any error, but can't see all the data. I am able to see only the headers.

    Even when I run the interface that loads data into a table, its operation without error, but no data entered...

    But the data is present in the source file...

    Can you please help me how to solve this problem...

    Hi Phanikanth,

    Thanks for your reply...

    I did the same thing that you suggested...

    In fact, I'm working on the ODI in UNIX environment. So I went for the record separator on UNIX option in the files of the data store tab and now its works well...

    in any case, once again thank you for your response...

    Thank you best regards &,.

    Vanina

  • Export to a CSV of tiara does not produce a CSV file!

    Seems strange but exporting to a CSV of tiara does not a CSV file.  Tiara produces a tab delimited file, which is obviously not a CSV file.  Tiara aid said even the 'CSV' file is a tab-delimited.  The software expects a comma as delimiter does not read the file properly created tiara.  It is possible to specify the correct delimiter?

    You can ask the decimal/separator, you need if you start exporting CSV from a script

    Call DataFileSave ("C:\temp\EXAMPLE.csv ', 'CSV')

    See also http://forums.ni.com/ni/board/message?board.id=60&message.id=8851&query.id=146775#M8851

  • Add-adgroupmember are ZERO values as possible when you import users from a csv file?

    I run the PS command against a csv file containing a list of ad groups followed by ADUsers below.

    The script works very well as long as all fields are filled out, there at - it a command / switch that ignores a value not as such?

    import-csv c:\admin\powershell\ADGroupMembers.csv | foreach {add-adgroupmember-identity $_.} {Ad group $_.member1, $_.member2, $_.member3, $_.member4, $_.member5}
    ContentsADGroup of CSV file member1, member2, Member3, Member4, Member5 AD-Test1, Minnie, Mickey, Donald, Daisy, goofy AD-Test1, Minnie, Mickey, Donald, Minnie, Mickey, Donald, Pete, AD-Test1, Spike

    Hello

    Your question is more complex than what is generally answered in the Microsoft Answers forums. It is better suited for Exchange Server on Technet. Please post your question in the Technet forums. You can follow the link to your question:

  • How can I extract the data from a csv file and insert it into an Oracle table? (UTL_FILE)

    Hi, please help me whit this query

    Im trying to extrate the data in a file csv and im using the ULT_FILE package
    I have this query that read the file and the first field, but if the field has a different length does not work as it should

    For example if I had this .csv file:

    1, book, laptop

    2, pen, Eraser

    3, notebook, paper

    And in the table, I had to insert like this

    ID descrption1 description2

    laptop 1 book

    Eraser pen 2

    paper laptop 3

    For now, I have this query, which displays only with DBMS:

    Declare

    -Variables

    Cadena VARCHAR2 (32767).

    Vfile UTL_FILE. TYPE_DE_FICHIER;

    Dato varchar2 (200);                                             -Date

    dato1 varchar2 (200);

    dato2 varchar2 (200);

    Identifier varchar2 (5): = ', ';                             -Identifier (en)

    v_ManejadorFichero UTL_FILE. TYPE_DE_FICHIER;     -For exceptions

    -Table variables

    I_STATUS GL_INTERFACE. % OF STATUS TYPE.

    I_LEDGER_ID GL_INTERFACE. TYPE % LEDGER_ID;

    I_USER_JE_SOURCE_NAME GL_INTERFACE. TYPE % USER_JE_SOURCE_NAME;

    I_ACCOUNTING_DATE GL_INTERFACE. TYPE % ACCOUNTING_DATE;

    I_PERIOD_NAME GL_INTERFACE. TYPE % PERIOD_NAME;

    I_CURRENCY_CODE GL_INTERFACE. CURRENCY_CODE % TYPE;

    I_DATE_CREATED GL_INTERFACE. DATE_CREATED % TYPE;

    I_CREATED_BY GL_INTERFACE. CREATED_BY % TYPE;

    I_ACTUAL_FLAG GL_INTERFACE. TYPE % ACTUAL_FLAG;

    I_CODE_COMBINATION_ID GL_INTERFACE. TYPE % CODE_COMBINATION_ID;

    I_ENTERED_DR GL_INTERFACE. TYPE % ENTERED_DR;

    I_ENTERED_CR GL_INTERFACE. TYPE % ENTERED_CR;

    I_ACCOUNTED_DR GL_INTERFACE. TYPE % ACCOUNTED_DR;

    I_ACCOUNTED_CR GL_INTERFACE. TYPE % ACCOUNTED_CR;

    I_TRANSACTION_DATE GL_INTERFACE. TRANSACTION_DATE % TYPE;

    I_REFERENCE1 GL_INTERFACE. REFERENCE1% TYPE;

    I_REFERENCE2 GL_INTERFACE. REFERENCE2% TYPE;

    I_REFERENCE3 GL_INTERFACE. REFERENCE3% TYPE;

    I_REFERENCE4 GL_INTERFACE. REFERENCE4% TYPE;

    I_REFERENCE5 GL_INTERFACE. REFERENCE5% TYPE;

    I_REFERENCE10 GL_INTERFACE. REFERENCE10% TYPE;

    I_GROUP_ID GL_INTERFACE. GROUP_ID % TYPE;

    BEGIN

    Vfile: = UTL_FILE. FOPEN ('CAPEX_ENVIO', 'comas.csv', 'R');

    loop

    UTL_FILE. GET_LINE(Vfile,Cadena,32767);

    dato1: = substr (cadena, instr(cadena, identificador,1,1)-1, instr(cadena, identificador,1,1)-1);

    dato2: = substr (cadena, instr (cadena, identifier, 1, 1) + 1, instr(cadena, identificador,3,1)-3);

    dbms_output.put_line (dato1);

    dbms_output.put_line (dato2);

    -The evidence

    -dbms_output.put_line (cadena);

    -dbms_output.put_line (substr (dato, 3, instr(dato, identificador,1,1)-1));

    -dbms_output.put_line (substr (dato, instr (dato, identifier, 1, 2) + 1, instr(dato, identificador,1,1)-1));

    -dbms_output.put_line (substr (cadena, 1, length (cadena)-1));

    end loop;

    UTL_FILE. FCLOSE (Vfile);

    -----------------------------------------------------------------------------------EXCEPTIONS------------------------------------------------------------------------------------------------------------------------------------------------------------

    EXCEPTION

    When no_data_found then

    dbms_output.put_line ('Todo Correcto');

    When utl_file.invalid_path then

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20060,'RUTA DEL ARCHIVO NULLIFIED: (');)

    WHEN UTL_FILE. INVALID_OPERATION THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR ('-20061,'EL ARCHIVO NO PUDO SER ABIERTO ");

    WHEN UTL_FILE. INVALID_FILEHANDLE THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20062, 'INVALIDO MANAGER');

    WHEN UTL_FILE. WRITE_ERROR THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20063, 'ESCRITURA ERROR');

    WHEN UTL_FILE. INVALID_MODE THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20064, 'MODO INVALIDO');

    WHEN UTL_FILE. INTERNAL_ERROR THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20065, 'ERROR INTERNO');

    WHEN UTL_FILE. READ_ERROR THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20066, 'LECTURA ERORR');

    WHEN UTL_FILE. FILE_OPEN THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR ('-20067,'EL ARCHIVO ARE ESTA ABIERTO ");

    WHEN UTL_FILE. THEN ACCESS_DENIED

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20068, 'REFUSED ACCESS');

    WHEN UTL_FILE. DELETE_FAILED THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20069, 'OPERACIÓN BORRADO FALLO');

    WHEN UTL_FILE. RENAME_FAILED THEN

    UTL_FILE. FCLOSE (V_ManejadorFichero);

    RAISE_APPLICATION_ERROR (-20070, 'OPERATION SOBREESCRITURA FALLO');

    END;

    Hello

    Try something like this:

    POS1: = INSTR (cadena, idntificador, 1, 1);
    POS2: = INSTR (cadena, idntificador, 1, 2);

    ID: = SUBSTR (cadena, 1, pos1 - 1);
    description1: = SUBSTR (cadena, pos1 + 1, (pos2 - pos1)-1);
    Description2: = SUBSTR (cadena, pos2 + 1);

    where pos1 and pos2 are numbers.

    Rather than use UTL_FILE, consider creating an external table.  You won't have to write any PL/SQL, and this means that you won't be tempted to write a bad article of EXCEPTION.

  • Analysis 1 row into two lines from a CSV file

    Hello

    Im trying to read a csv of the external table data in a table target.

    the problem is, I have a few lines, two names and names separated by spaces zwo (aspect ID2 and ID4)

    the csv data have this format:

    Source CSV file

    ID1 | "" Max Miller ' | "Lyonerstr 99" | " "1000" | " "" Berlin "| "' The Germany.

    ID2. «Hans Meyer Heidi Meyer «|» "Lyonerstr 100" | " "1000" | " "" Berlin "| "' The Germany.

    ID3 | "" Stefan Tek | "Lyonerstr 200" | " "1000" | " "" Berlin "| "' The Germany.

    ID4. ' José María Acero Acero ' |. ' "" Abcstr 111 | "2000" | " "" Hamburg ". "' The Germany.

    Target table

    ID1 | Max Miller | 99 Lyonerstr | 1000 | Berlin | Germany

    ID2. Hans Meyer | Lyonerstr 100 | 1000 | Berlin | Germany

    ID2. Heidi Meyer | Lyonerstr 100 | 1000 | Berlin | Germany

    ID3 | Stefan Tek | Lyonerstr 200 | 1000 | Berlin | Germany

    ID4. José Acero | Abcstr 111 | 2000. Hamburg | Germany

    ID4. Maria Acero | Abcstr 111 | 2000 "|" Hamburg | Germany

    Thank you very much.

    with

    external_table as

    (select 'ID1' u_id, f_l_name 'Max Miller', ' Lyonerstr 'address, zip ' 1000' 99, "Berlin" city, country "Germany" in double union ")

    Select 'ID2', 'Hans Meyer Heidi Meyer', ' Lyonerstr 100 ', ' 1000', 'Berlin', 'Germany' of the dual union all.

    Select 'ID3', "Stefan Tek", "Lyonerstr 200 ', ' 1000', 'Berlin', 'Germany' of the dual union all.

    Select "ID4", 'José Acero Acero Maria', ' Abcstr 111 ', ' 2000', 'Hamburg', 'Germany' from dual.

    )

    Select u_id, f_l_name, address, zip, city, country

    from (select u_id,

    -case when instr (f_l_name,' ') > 0

    so to case when level = 1

    then substr (f_l_name, 1, instr (f_l_name,' ')-1)

    of another substr (f_l_name, instr (f_l_name,' ') + 2)

    end

    another case when level = 1

    then f_l_name

    end

    end f_l_name

    address, zip, city, country

    of external_table

    connect by level<=>

    and prior u_id = u_id

    and the previous address = address

    zip and rar = prior

    and prior city = city

    and prior country = country

    and prior sys_guid() is not null

    )

    where f_l_name is not null

    U_ID F_L_NAME ADDRESS ZIP CITY COUNTRY
    ID1 Max Miller 99 Lyonerstr 1000 Berlin Germany
    ID2 Hans Meyer Lyonerstr 100 1000 Berlin Germany
    ID2 Heidi Meyer Lyonerstr 100 1000 Berlin Germany
    ID3 Stefan Tek Lyonerstr 200 1000 Berlin Germany
    ID4 José Acero Abcstr 111 2000 Hamburg Germany
    ID4 Maria Acero Abcstr 111 2000 Hamburg Germany

    Concerning

    Etbin

  • Numbers does not correctly Import CSV files

    Using numbers v 3.6.2 under OS X El Capitan v 10.11.6 CSV files are not imported correctly. A file with 8 columns appears in number with the first two columns containing data of 7 columns and the second containing the remaining column data. This does not happen with numbers v 2.1 I have on another Mac OS x 10.6.8

    Well, I know one answer would be to always download on the Mac of the latter, but I'm trying to stop using it, once all the data I need is available on the other, later, Mac.

    If Apple could provide a program that does work properly, why on Earth can't a version later do the same thing, it is a fundamental requirement and I don't want to play anywhere to try to sort out just that it works as it should.

    might be useful to provide an example of the file, or, perhaps to share data... Nobody here can change the numbers, we are all of the users like you.  The only options that we are thus solutions using what is already there, or work around issues by other means.

    You can still provide feedback to post using the menu item "numbers > provide Feedback numbers.

  • OracleTextSearch doesn't work is not for the CSV file

    Hi all

    I use Oracle Webcenter content with Version: 11.1.1.9.0 - 2015-04-14 07:19:29Z - r126792 (Build: 7.3.5.185).

    I configured using OracleTextSearch below of setting in the config.cfg file

    SearchIndexerEngineName = OracleTextSearch

    I can search text for the file types such as text, excel, docx, pdf, but I'm not able to do text search the csv file.

    Please suggest any alternative to enable text search the csv file.

    Please find the below the example of csv file.

    "Id","CurrencyIsoCode","City","Status"
    "1","GBP","","Activated"
    "2","GBP","Reading","Expired"
    "3","GBP","Reading","Expired"
    "4","GBP","Brighton","Activated"
    "5","USD",""Atlanta"""","" Georgia""","Expired"
    "6","GBP","Reading","Expired"
    "7","GBP","Bristol","Activated"
    "8","GBP","Chicago","Expired"
    "9","GBP","Chicago","Expired"
    "10","GBP","Chicago","Expired"
    "11","GBP","Chicago","Expired"
    "12","GBP","Reading","Activated"
    "13","GBP","London","Activated"
    "14","GBP","Reading","Expired"
    "15","GBP","Singapore 068893","Activated"
    "16","USD","Overland Park","Activated"
    "17","GBP","London","Activated"
    "18","GBP","Hangzhou","Activated"
    "19","GBP","Southampton","Activated"
    

    Do something:

    Edit config.cfg and add below variable:

    TextIndexerFilterFormats = csv

    After that, save and exit. Restart the server.

    Checking in a new file csv and do a word search.

    It should work let me know again, you have a question or not

  • Delete snapshots by using data from a csv file

    I have a csv file that was exported in the form of:

    VM

    SERV1

    serv2

    SERV3

    (The file name is snaps4.csv)

    I want to delete the associated clichés a vm in this file csv; However I can't get anything to work.  Is the closest I've come by manually removing the header in the csv file (i.e. VM) then using the get-content command.

    $vms = get-Content C:\scripts\Output\snaps.csv

    Get-Snapshot - $vms vm | Remove-Snapshot - RemoveChildren-confirm: $false

    The above command works, but I have to remove the header first (which I am fine with, if I do this, but I'm trying to automate this process for people of our operations and have a manual step is not ideal).

    Someone help me?  I know I'm missing just something simple here, but can't understand it.


    Have you tried it?

    Import-Csv C:\scripts\Output\snaps.csv | %{

    Get-Snapshot - vm $_. VM | Remove-Snapshot - RemoveChildren-confirm: $false

    }

  • Helps to replace a string in a txt file with a string from a csv file

    Hi all

    I worked on the script following since a few days now, determined to exhaust my own knowledge before asking for help... Unfortunately, it didn't take very long to exhaust my knowledge :-(

    Basically, I need to replace a value in a single file (raw.txt) with the value of another file (userids.csv) when a variable is. Then I released the results of a third file.

    To do this, I divided the "raw" file into variables using the ',' as the separator, the problem is that some variables are intentionally empty, where the fi $variable = "statements.

    It is currently what I want to do but only when the userids.csv file contains a single line. It is obviously because of the foreach ($user in import)... What I need to figure out is how to loop through the file raw.txt, text replacement when a variable in the user ID file is the text in raw.txt... I hope that makes sense?

    The user ID file is in the following format - user, service, Dept that can contain dozens of lines

    I would appreciate any pointers :-)

    See you soon

    # Treatment
    $importraw = get-content i:\raw.txt
    $import = import-csv i:\userids.csv-en-tete UserAccount, functional, Dept

    {foreach ($user in $import)
    $useraccount = $user. UserAccount
    $userfunction = $user. Functional
    $userdept = $user. Dept
    {foreach ($line in $importraw)
    $first, $second, $third, $fourth, $fifth, $sixth, $seventh, $eighth, $ninth = $line - split(",")
    $linesproc = $linesproc + 1
    If ($sixth - eq ") {}
    $temp6 = "6TEMP".
    Write-Host "field Null detected - assigning temporary value:"$temp6 ".
    $sixth = $temp6 # the assignment of a temporary value so that - statement to replace later works
    }
    If ($seventh - eq ") {}
    $temp7 = "7TEMP".
    Write-Host "field Null detected - assigning temporary value:"$temp7 ".
    $seventh = $temp7 # the assignment of a temporary value so that - statement to replace later works
    }
    If ($fifth - eq $user.) UserAccount) {}
    $line - $seventh, replace $user. Dept | Add content i:\Output.txt
    }
    else {}
    $line - $seventh, replace "/ / customer. Add content i:\Output.txt
    }
    }

    }

    Try the attached version.

    The problem, in my opinion, was in nested ForEach loops.

    Instead I've implemented with a lookup table

  • ColdFusion does not recognize the CSV files

    I have a local version of ColdFusion developer 10 on my local machine.

    I use ColdFusion to create a CSV file and save it on the server and create a link by using the following code.

    Browsers open the CSV file in a text file and I use link right click to save target, my Windows Explorer save a html file.

    It works on the other server, using the same code.

    I want to know are there any which configuration to set ColdFusion recognize the format of CSV file when saving the target and he lets browsers know how to use Excel to open CSV files.

    Your information and help is much appreciated,

    Kind regards

    Mouses,

    < Cfset thisPath = ExpandPath("*.*") >

    < cfset f_dir = GetDirectoryFromPath (thisPath) >

    < cfset what = "#dateformat (now (), 'mmddyy') ##timeformat (now (), 'hhmm') # .csv" >

    < cffile action = "WRITE" line = "" f_dir # # how # ""

    output = "MtField1, MyField2" addnewline = "Yes" >

    < cfloop query = "#myQuery #" >

    < cffile action = "APPEND" file = "" f_dir # # how # ""

    output = "Field1 # #, #Field2 # ' addnewline ="Yes">

    < / cfloop >

    < cfset lelien = "MyServer/report /" & what > »»

    < cfoutput > < a href = "#theLink #" > here is the file < /a > < / cfoutput >

    < br >

    There are several different themes play. I have them separate to make the code more manageable and scalable.

    Using your code, I've implemented the following quick test. It uses the embedded, cfartgallery data source, you have also.

    I placed the 3 files in the same directory. But this is usually unnecessary because you can pass the path of the directory as a variable.

    fileWrter.cfm

    Select the city, email

    artists

    Here is the file

    fileDownload.cfm

    You must first create the download file.

    File.CFC

  • Script to remove commas from a csv file

    Hi all

    I have the following output to a csv file:

    VM, VI-SVC-VM014, 0000ed, 0000ee,
    VM, VI-SVC-VM103, 0000f3, 0000f2,
    VM, VI-SVC-VM104, 0000f6, 0000f6,
    LDEV, SVC PROD Cluster01, 0000e2, 0x04
    LDEV, SVC PROD Cluster01, 0000de, 0x00
    LDEV, SVC PROD Cluster01, 0000df, 0x01
    LDEV, SVC PROD Cluster01, 0000e2, 0x04

    I'm looking to remove commas end so that the output looks like this:

    VM, VI-SVC-VM014, 0000ed, 0000ee
    VM, VI-SVC-VM103, 0000f3, 0000f2
    VM, VI-SVC-VM104, 0000f6, 0000f6
    LDEV, SVC PROD Cluster01, 0000e2, 0x04
    LDEV, SVC PROD Cluster01, 0000de, 0x00
    LDEV, SVC PROD Cluster01, 0000df, 0x01
    LDEV, SVC PROD Cluster01, 0000e2, 0x04

    The column length for each line (given the number of objects), modular in this output. By this, I mean that some outputs will be like this:

    VM, VI-SVC-VM014, 0000ed, 0000ee, 0000ef
    VM, VI-SVC-VM104, 0000f6, 0000f6

    VI-SVC-VM104, 0000f6, 0000f5, 0000f7, VM, 0000f4
    LDEV, SVC PROD Cluster01, 0000e2, 0x04
    LDEV, SVC PROD Cluster01, 0000de, 0x00

    Is it possible to import the data (get-content), delete all non alpha-numeric end of line? Maybe there is an easier way.

    Thank you

    Rob.

    Try something like

    Get-Content file.csv | %{
      $_.TrimEnd(',')
    } | Set-Content newfile.csv
    
  • Externalize strings not generating a CSV file translations

    Hi gurus,

    I need to do outsourcing of presentation layer, so doing it, I activated by clikcing to outsource the table on the area in question and after I have go to tools-> utilities-> externalize strings-> run

    the csv file, it creates is having only the English, I don't get the translations to other languages in the csv file. I m hit there.

    Pls help!


    Thank you
    Anthony

    Hi Anthony,.

    The generated CSV file will be in English only. If you try to just change in OBIEE Presenatation layer to another language, you need to perform operations of addition after the export of the CSV file on your machine.

    Look at this article: (I think that's what you're looking for);

    http://oraclebizint.WordPress.com/2007/10/29/Oracle-BI-EE-101332-presentation-layer-translation/

    Kind regards
    -Laurence.

  • SQLLDR - load only selected columns from a csv file

    Hello

    Can we charge only a few selected columns of a big (INFILE) CSV using SLLDR?

    Thank you in Advcance
    Kapil

    Hello

    You can take a look at this. It may be useful

    DECLARE
    l_start_file_name varchar2(50) := 'invoice_excel_c';
    l_end_file_name varchar2(50)   := '.csv';
    l_file_name varchar2(100);
    l_sql varchar(32767);
    refresh_cdc varchar2(5) := '00304';
    
    begin
    
    l_file_name := l_start_file_name||REFRESH_CDC||l_end_file_name;
    
     l_sql :=
    'create table exter_invoice_excel               '||
    '(                                              '||
    'Product_name                varchar2(255),     '||
    'Net_Sales                   Number,            '||
    'Net_Adjustments             Number,            '||
    'Cancels_Count               Number,            '||
    'Cancels_Amount              Number,            '||
    'Cashes_Count                Number,            '||
    'Cashes_Amount               Number,            '||
    'Claims_Count                Number,            '||
    'Claims_Amount               Number,            '||
    'Returns_Count               Number,            '||
    'Returns_Amount              Number,            '||
    'Free_Prize_Count            Number,            '||
    'Free_Prize_Amount           Number,            '||
    'Free_Promo_Count            Number,            '||
    'Free_Promo_Amount           Number,            '||
    'Promo_Credit_Count          Number,            '||
    'Promo_Credit_Amount         Number,            '||
    'Return_Commission           Number,            '||
    'Net_Discounts               Number,            '||
    'Total_Fees                  Number,            '||
    'Sales_Commission            Number,            '||
    'Cash_Commission             Number,            '||
    'Tkt_Charge                  Number,            '||
    'Subscription_Commission     Number,            '||
    'Interim_Sweeps              Number,            '||
    'Net_Due                     Number,            '||
    'Retailer_id                 Number,            '||
    'Retailer_Name               varchar2(255)      '||
    ')                                              '||
                  'ORGANIZATION EXTERNAL (                 '||
                  ' TYPE oracle_loader                     '||
                   ' DEFAULT DIRECTORY GTECHFILES          '||
                   ' ACCESS PARAMETERS (                   '||
                   '   RECORDS DELIMITED BY NEWLINE        '||
                   '      CHARACTERSET WE8MSWIN1252        '||
                   '   BADFILE ''invoice_excel.bad''       '||
                   '   DISCARDFILE ''invoice_excel.dis''   '||
                   '   LOGFILE ''invoice_excel.log''       '||
                   '   FIELDS TERMINATED BY '','' RTRIM      '||
                   '  MISSING FIELD VALUES ARE NULL        '||
                   '   )                                   '||
                   '   LOCATION ('''||l_file_name||''')    '||
                   ' )                                     '||
                   'PARALLEL                               '||
                   'REJECT LIMIT UNLIMITED ';
    
            execute immediate l_sql;
    
          l_sql:=  'INSERT INTO INVOICE_EXCEL_TEMP                '||
                      ' ( Product_name               ,            '||
                      '  Net_Sales                   ,            '||
                      '  Net_Adjustments             ,            '||
                      '  Cancels_Count               ,            '||
                      '  Cancels_Amount              ,            '||
                      '  Cashes_Count                ,            '||
                      '  Cashes_Amount               ,            '||
                      '  Claims_Count                ,            '||
                      '  Claims_Amount               ,            '||
                      '  Returns_Count               ,            '||
                      '  Returns_Amount              ,            '||
                      '  Free_Prize_Count            ,            '||
                      '  Free_Prize_Amount           ,            '||
                      '  Free_Promo_Count            ,            '||
                      '  Free_Promo_Amount           ,            '||
                      '  Promo_Credit_Count          ,            '||
                      '  Promo_Credit_Amount         ,            '||
                      '  Return_Commission           ,            '||
                      '  Net_Discounts               ,            '||
                      '  Total_Fees                  ,            '||
                      '  Sales_Commission            ,            '||
                      '  Cash_Commission             ,            '||
                      '  Tkt_Charge                  ,            '||
                      '  Subscription_Commission     ,            '||
                      '  Interim_Sweeps              ,            '||
                      '  Net_Due                     ,            '||
                      '  Retailer_id                 ,            '||
                      '  Retailer_Name               ,            '||
                      '  SOURCE                      ,            '||
                      '  INSERTED_DATE               ,            '||
                      '  CDC                         ,            '||
                      '  UPLOADED                                 '||
                      ')                                          '||
    '              select                                         '||
                      '  ltrim(rtrim(product_name)) ,             '||
                      '  Net_Sales                   ,            '||
                      '  Net_Adjustments             ,            '||
                      '  Cancels_Count               ,            '||
                      '  Cancels_Amount              ,            '||
                      '  Cashes_Count                ,            '||
                      '  Cashes_Amount               ,            '||
                      '  Claims_Count                ,            '||
                      '  Claims_Amount               ,            '||
                      '  Returns_Count               ,            '||
                      '  Returns_Amount              ,            '||
                      '  Free_Prize_Count            ,            '||
                      '  Free_Prize_Amount           ,            '||
                      '  Free_Promo_Count            ,            '||
                      '  Free_Promo_Amount           ,            '||
                      '  Promo_Credit_Count          ,            '||
                      '  Promo_Credit_Amount         ,            '||
                      '  Return_Commission           ,            '||
                      '  Net_Discounts               ,            '||
                      '  Total_Fees                  ,            '||
                      '  Sales_Commission            ,            '||
                      '  Cash_Commission             ,            '||
                      '  Tkt_Charge                  ,            '||
                      '  Subscription_Commission     ,            '||
                      '  Interim_Sweeps              ,            '||
                      '  Net_Due                     ,            '||
                      '  Retailer_id                 ,            '||
                      '  Retailer_Name               ,            '||
                      ' '''||l_file_name                           ||''', '||
                      '   sysdate                    ,            '||
                      ' '''||    REFRESH_CDC                       ||''', '||
                      '''N'''                                      ||
                      'FROM                                       '||
                      'exter_invoice_excel                        '||
                      'WHERE retailer_id is not null and ltrim(rtrim(product_name)) in (''Loto'',''Inst Tk'')';
    
            execute immediate l_sql;    
    
            l_sql :='drop table exter_invoice_excel';
            execute immediate l_sql;
    
    exception
    when others then
    rollback;
    l_sql :='drop table exter_invoice_excel';
    execute immediate l_sql;
    debug_message('INVOICE_EXCEL_UPLOAD/'||REFRESH_CDC,'Unexpected Error '||sqlerrm);
    END;
    
    --truncate table invoice_excel_temp
    

    Thank you

Maybe you are looking for

  • Payment

    I have to pay for this music to download if I use Apple music?

  • Apple tv bluethoot matching

    I am not able to pair my esrphone bluethoot with my 2nd generation or 3rd of apple tv. the pair of headphones with the ipad and the ipod, but not with apple tv. is it possible to do?

  • 4620 ink - black Office jet does not print

    Even if the ink level was at 20% level, not printing black ink. I cleaned the print head, but no resolution. Then, I bought a new HP cartridge at Staples and replaced the old. Still not printing black. Other colors ok. Went through a few hours of cle

  • How to get the system tray at the bottom of the monitor

    I found that the system tray icons had moved to the right side of the monitor. How do I get back them down

  • E-mail TMSXE 3.1 templates

    I just upgraded from TMSXE 3.0 3.1.  This was due to the incompatibility between the e-mail templates - we wanted people to get the same notifications if they were reserved by TMS or via outlook and the release notes said: "New templates of notificat