NFS mounts persist?
Hi all
I've updated to Win 7 Enterprise for NFS support.
I was able to map two NFS shares successfully.
However, upon restart, the actions are lost. Is it possible (I don't see one in the snap-snap-in mmc NFS) to indicate that they be reconnected at logon. I could use a login script in local GPO to do this, but I was wondering if there was a "better way" (tm).
A tip for those who struggle with the NFS client: use showmount to see if Windows can even see the actions. I had to set the firewall issues, then showmount showed stocks and I was able to login and navigate through the:
Server EI C:\Users\owner>showmount
Export the server list:
/nfs4exports 192.168.x.0/24
RAID 192.168.x.0/24/nfs4exports.
mythstorage 192.168.x.0/24/nfs4exports.
C:\Users\owner>mount server: / nfs4exports/z: raid
z: is now connected to the server: / nfs4exports/raid
The command completed successfully.
C:\Users\owner >
Found the answer on the technet site (it's sort of, a persistent systemic mount is not possible, but using a user login script is possible.)
Basically NFS connections are stored on a basis by user.
"NFS mounted drive letters are specific session so running this in a script to start most likely will not work." The service would probably need to mount its own drive letters from a script that runs in the context of service. We have specific guidelines, that it is not recommended to http://support.microsoft.com/kb/180362. "
Tags: Windows
Similar Questions
-
T2p scripts fail on NFS mounts
Hello IDM gurus!
I'm trying to implement a multi data center (MDC) implementation of OAM 11gR2Ps3. 1 data center is in place and functioning as expected. We try to use T2P scripts described in the documentation of the OAM following MDC to clone the binary (copyBinary.sh) but you see the error message
Error message: 1
November 17, 2015 18:46:48 - ERROR - CLONE-20435 some Oracle homes are excluded during the copy operation.
November 17, 2015 18:46:48 - CAUSE - CLONE-20435 following Oracle homes have been excluded during the copy operation.
[/ data/Oracle/OAM/.snapshot/daily.2015-11-13_0010/oracle_common, data/Oracle/OAM/.snapshot/daily.2015-11-13_0010/oracle_bip, data/Oracle/OAM/.snapshot/daily.2015-11-13_0010/Oracle_IDM1, data/Oracle/OAM/.snapshot/daily.2015-11-14_0010/oracle_common, data/Oracle/OAM/.snapshot/daily.2015-11-14_0010/oracle_bip, data/Oracle/OAM/.snapshot/daily.2015-11-14_0010/Oracle_IDM1, data/Oracle/OAM/.snapshot/daily.2015-11-15_0010/oracle_common, /data/Oracle/OAM/.snapshot/daily.2015-11-15_0010/oracle_bip] data/Oracle/OAM/.snapshot/daily.2015-11-15_0010/Oracle_IDM1, data/Oracle/OAM/.snapshot/daily.2015-11-16_0010/oracle_common, data/Oracle/OAM/.snapshot/daily.2015-11-16_0010/oracle_bip, data/Oracle/OAM/.snapshot/daily.2015-11-16_0010/Oracle_IDM1, data/Oracle/OAM/.snapshot/daily.2015-11-17_0010/oracle_common, data/Oracle/OAM/.snapshot/daily.2015-11-17_0010/oracle_bip, data/Oracle/OAM/.snapshot/daily.2015-11-17_0010/Oracle_IDM1, /data/Oracle/OAM/.snapshot/daily.2015-11-11_0010/oracle_common /data/Oracle/OAM/.snapshot/daily.2015-11-11_0010/oracle_bip [, data/Oracle/OAM/.snapshot/daily.2015-11-11_0010/Oracle_IDM1, data/Oracle/OAM/.snapshot/daily.2015-11-12_0010/oracle_common, data/Oracle/OAM/.snapshot/daily.2015-11-12_0010/oracle_bip, data/Oracle/OAM/.snapshot/daily.2015-11-12_0010/Oracle_IDM1], and the possible causes are:
1. all Oracle homes were not registered with an OraInventory.
2 if all the Oracle homes have been saved with a unique custom OraInventory, pointer file corresponding inventory received no T2P operation.
3 canonical path to the Oracle Home recorded with OraInventory was not child Middleware House.
November 17, 2015 18:46:48 - ACTION - CLONE - 20435. be sure that all causes possible, listed in the CAUSE section are supported.
.Snapshot in the error message file is a NFS function because it mounts are actually NFS volumes.
Are the T2P scripts supported on NFS mounts? If so, how to solve this problem? Would appreciate your comments on how we can solve this problem
Concerning
Shiva
The issue was due to wrong settings for the frames. This has been fixed by the storage team. T2P scripts were always initially and after that correction of the Storage mounts, they ran successfully. A restoration was necessary, and RCU patterns had to be recreated on the clone
-
ESXi 4.1 - blades HP and NFS mounts
Hello
I'm having a strange problem with NFS access
We currently use HP Blades BL680c G5 servers, in C7000 enclosures. Configuration of the switch, Cisco (3120 x) are the same for all ports / ESXs.
I use VMNIC0 and VMNIC1 in etherchannel / Hash IP access to the warehouses of data NFS on NetApp filers.
I have now two servers beside I want to reverse for test reasons. These two servers are moving instead of the other.
Once restarted, these two hosts ar sees no no more data warehouses. They appear as inactive. I can't reconnect them (VI client or esxcfg-nas - r)
vmkping for Netapp file servers works more also.
As soon as I put these 2 servers to their original square, everything is back to normal.
On the side of the cisco switch ports are still configured the same. None of them are closed.
There I'm missing something, but don't know where to look.
Thanks in advance
Hi Hernandez,
If you're repatching hosts or they always use the same network cables as before (perhaps a change of IP address, or a rule on the host of the NFS export)
don't forget, you'll have policies created on the NFS mounts 'export' these fixtures for the names of specific hosts or IPs - if you're repatching and change IP etc you can cause this export invalidation,
-
I have two questions about how the NFS mount points is in fact from the point of view NFS servers.
If the mount point of the 'outside' is considered as for example / ISO, would this mean
1. that the directory is named "iso" or could it be a real name on the local file system on the NAS Server? (As in the names of SMB share, could be different from the name of folder.)
2. that the folder is actually located directly on the root of the local file system, which is/ISO on the server, or would it be for example/dir1/dir2/iso and always to the outside as appearance / ISO? (As in the SMB, where 'share' creates a new root somewhere on the filesystem that is exposed to the outside.)
ricnob wrote:
Thank you! It would be a typical NFS-Unix/Linux server?
Yes.
So perhaps it is not defined in the standard how this should be done, but can vary between implementations? (* nix of course most common.)
Possible, but does MS ever meet standards? Even their own?
RFC1094/NFS:
"Different operating systems may have restrictions on the depth of the tree or used names, but also using a syntax different tor represent the"path", which is the concatenation of all the"components"(file and directory names) in the name...". Some operating systems provide a 'mount' operation to make all file systems appear as a single tree, while others maintain a 'forest' of file systems... NFS looking for a component of a path at the same time. It may not be obvious why he did not only take the name of way together, stroll in the directories and return a file handle when it is carried out. There are many good reasons not to do so. First, paths need separators between items in the directory, and different operating systems use different separators. We could define a representation of network Standard access path, but then each path must be analyzed and converted at each end. "
AWo
VCP 3 & 4
\[:o]===\[o:]
= You want to have this ad as a ringtone on your mobile phone? =
= Send 'Assignment' to 911 for only $999999,99! =
-
maximum size of the NFS mounts
Hello
I would like to know what is the maximum size for mounting NFS into ESX (i) 4? If I need more than 2 TB, I need to use extensions iSCSI volumes. Is it the same thing with NFS or I can come up and use for example a 4 TB NFS share?
Thank you
Marcus
NFS mounts have No. limit. It is also great that you need to make. ESX can mount NFS, that's all. The ONLY limitation for ESX is VMFS, because ESX creates / manages VMFS volumes. NFS mounts are managed by OTHER systems, they have their own limits.
NFS is just a huge disk hard as far as ESX is concerned.
-
How to start vmware after NFS mounts
OK, so now I have solved the mystery of editing. I'm looking for the gurus of linux on how to change the sequence of boot for mounts NFS are made prior to the VMware service.
I use Centos 5.3 on the host.
Phil
Hello
You must configure vmware to start after nfs and network.
A simple example makes it easier to understand:
LS /etc/rc3.d/S*network
/etc/rc3.d/S10network
LS /etc/rc3.d/S*nfs
/etc/rc3.d/S11nfs
LS /etc/rc3.d/S*vmware
/etc/rc3.d/S05vmware
In this case, it may be like this:
RM /etc/rc3.d/S05vmware
ln-s /etc/init.d/vmware /etc/rc3.d/S20vmware
Today, vmware always will begin after the nfs mounts.
A saludo/best regards,.
Pablo
Please consider providing any useful answer. Thank you!! - Por favor considered premiar las useful responses. MUCHAS gracias!
-
I just bought a Qnap NAS box that I want to share with 2 VM hosts. I did a little research on NFS vs iSCSI and it seems that NFS has some pretty good advantage over iSCSI. Especially when she is on a network to shared resources.
The hosts are the two Centos 5.3.
I wonder what would be the optimum settings (options) for the NFS mount? I found a couple of some old site, but I wonder what people here use? Here's a setting string that I found:
nfsvers = 3, tcp, timeo = 600, TRANS = 2, rsize = 32768, wsize = 32768, hard, intr, noatime
I'm heading to the coast to look at the man page for NFS now to see what these all mean...
Thank you
Phil
Here's a good explanation of this thing:
The nolock option prevents the sharing of file locking
information between the NFS server and the NFS client. The server is
unaware of the file locks on this client and vice versa.
Using
the nolock option is required if the NFS server has its NFS file lock
the feature in a State broken or remained dead letter, but she works between
all versions of NFS.
If another host is accessing the same files that
your app on this host, there may be problems if file locking
information sharing is disabled.
Failure to maintain good
blocking a write operation to the same host with a read operation on
another may cause the reader to have incomplete or inconsistent data
(reads a structure of data/record/line that is written only in part).
A
locking failure between two writers is likely to have data
loss/corruption, such as the later writing replaces the previous one. The
changes made by the previous write operation may be lost.
-
-
Try adding that another NFS mount to a 3.5 server. I have an existing one on a Windows 2003 server that fits very well. The error I get when creating is:
& gt; Configuring the host error: cannot open the volume: / vmfs/volumes/f33bc09a-59b57f41 & lt;
Any ideas? It is also a 2003 server, but different from the box.
Hello
Looks like the esx host IP or subnet does not have access to the NFS share.
Check the properties of the NFS server.
vExpert 2009
-
NFS ODA persists even after the update applied
Good morning guys,
We have improved our ESXi Hypervisor from 5.1 to 5.5 U1 last weekend. I have read on the NFS ODA bug and downloaded the update scheduled to apply last nigth.
~ # software esxcli vib install d ' "/ vmfs/volumes/4f27d555-e55efb08-0da4-d4ae52723fbc /ESXi550 - 201407001.zip".
Result of the installation
Message: The update completed successfully, but the system must be restarted for the changes to be effective.
Restart required: true
VIBs installé : VMware_bootbank_esx-base_5.5.0-1.28.1892794, VMware_bootbank_lsi-mr3_0.255.03.01-2vmw.550.1.16.1746018, VMware_bootbank_lsi-msgpt3_00.255.03.03-1vmw.550.1.15.1623387, VMware_bootbank_misc-drivers_5.5.0-1.28.1892794, VMware_bootbank_mtip32xx-native_3.3.4-1vmw.550.1.15.1623387, VMware_bootbank_net-e1000e_1.1.2-4vmw.550.1.15.1623387, VMware_bootbank_net-igb_5.0.5.1.1-1vmw.550.1.15.1623387, VMware_bootbank_net-tg3_3.123c.v55.5-1vmw.550.1.28.1892794, VMware_bootbank_rste_2.0.2.0088-4vmw.550.1.15.1623387, VMware_bootbank_sata-ahci_3.0-18vmw.550.1.15.1623387, VMware_bootbank_scsi-megaraid-sas_5.34-9vmw.550.1.28.1892794, VMware_bootbank_scsi-mpt2sas_14.00.00.00-3vmw.550.1.15.1623387-VMware_locker_tools-light_5.5.0-1.28.1892794
VIBs supprimés : VMware_bootbank_esx-base_5.5.0-0.0.1331820, VMware_bootbank_lsi-mr3_0.255.03.01-1vmw.550.0.0.1331820, VMware_bootbank_lsi-msgpt3_00.255.03.03-1vmw.550.0.0.1331820, VMware_bootbank_misc-drivers_5.5.0-0.0.1331820, VMware_bootbank_mtip32xx-native_3.3.4-1vmw.550.0.0.1331820, VMware_bootbank_net-e1000e_1.1.2-4vmw.550.0.0.1331820, VMware_bootbank_net-igb_2.1.11.1-4vmw.550.0.0.1331820, VMware_bootbank_net-tg3_3.123c.v55.5-1vmw.550.0.0.1331820, VMware_bootbank_rste_2.0.2.0088-4vmw.550.0.0.1331820, VMware_bootbank_sata-ahci_3.0-17vmw.550.0.0.1331820, VMware_bootbank_scsi-megaraid-sas_5.34-9vmw.550.0.0.1331820, VMware_bootbank_scsi-mpt2sas_14.00.00.00-3vmw.550.0.0.1331820-VMware_locker_tools-light_5.5.0-0.0.1331820
After the reboot of the host and reassemble our NFS (used for backup VMs using ghettoVCB.sh), ODA issues are persistent that I checked on the /var/log/vobd.log this morning:
2014-07 - 10 T 11: 00:58.757Z: [APDCorrelator] 46073957136us: [vob.storage.apd.start] device or file with identifier [a643d5cd-6c9ea269] system has entered the State of all the paths down.
2014-07-10 T 11: 00:58.757Z: [APDCorrelator] 46073957584us: [esx.problem.storage.apd.start] device or filesystem with identifier [a643d5cd-6c9ea269] has entered the State of all the paths downwards.
2014-07-10 T 11: 02:45.898Z: no correlator for vob.vmfs.nfs.server.disconnect
2014-07-10 T 11: 02:45.898Z: [vmfsCorrelator] 46181098984us: [esx.problem.vmfs.nfs.server.disconnect] 192.168.100.83/mnt/HD/HD_a2/VMBACKUP a643d5cd-6c9ea269-0000-000000000000 D-Link
2014-07-10 T 11: 03:18.758Z: [APDCorrelator] 46213958621us: [vob.storage.apd.timeout] device or filesystem with identifier [a643d5cd-6c9ea269] has entered the State of all the paths down Timeout after being in the State of all the paths downwards for 140 seconds. E/s will now be quickly failed.
2014-07-10 T 11: 03:18.758Z: [APDCorrelator] 46213959037us: [esx.problem.storage.apd.timeout] device or filesystem with identifier [a643d5cd-6c9ea269] has entered the State of all the paths down Timeout after being in the State of all the paths downwards for 140 seconds. E/s will now be quickly failed.
2014-07-10 T 11: 07:40.806Z: [APDCorrelator] 46476006085us: [vob.storage.apd.exit] device or filesystem with identifier [a643d5cd-6c9ea269] left the State of all the paths downwards.
2014-07-10 T 11: 07:40.806Z: no correlator for vob.vmfs.nfs.server.restored
2014-07-10 T 11: 07:40.806Z: [APDCorrelator] 46476006585us: [esx.clear.storage.apd.exit] device or filesystem with identifier [a643d5cd-6c9ea269] left the State of all the paths downwards.
2014-07-10 T 11: 07:40.806Z: [vmfsCorrelator] 46476006474us: [esx.problem.vmfs.nfs.server.restored] 192.168.100.83/mnt/HD/HD_a2/VMBACKUP a643d5cd-6c9ea269-0000-000000000000 D-Link
So, anyone has an idea about this?
Thanks in advance.
Hey fabio_brizzolla,
Well the first thing I found was that the D - Link DNS-320 is not on the hardware compatibility list VMwares, however some other versions are, so unless I missed something I couldn't find it, which means that there could be unexpected behavior of the installation.
Having said that, your other editing that are the same, they all work well on a 5.0 update x is just installing 5.5 that you are having problems with?
-
Unalterable audit NFS mounted data warehouses
Hi all
I searched for a cmdlet or function that returns if an NFS volume is mounted read/write or read only on a node. The use case for this looping through a vCenter to ensure volumes that need to be mounted read-only were not accidentally mounted read/write during a reconstruction of node. I haven't found something that returns this value.
Is there something available in PowerCLI who will provide this information?
Thank you
Jason
Try something like this
foreach($ds in (Get-Datastore | where {$_.Type -eq "NFS"})){ $ds.ExtensionData.Host | Select @{N="DS";E={$ds.Name}}, @{N="RemoteHost";E={$ds.RemoteHost}}, @{N="RemotePath";E={$ds.RemotePath}}, @{N="ESXi";E={Get-view $_.Key -Property Name | Select -ExpandProperty Name}}, @{N="AccessMode";E={$_.MountInfo.AccessMode}}}
If you want to only see those read-only, you can add a Where clause
-
Windows Server 2008 R2 NFS mounted data store is read only
I've done this dozens of times in the past, but not with this exact combination successfully:
ESX4 U1
Windows 2008 R2 Server Datacenter edition
NFS Server installed. TCP Version 3.
The normal configuration that worked in the past:
NTFS security - total control on ANONYMOUS logon, everyone control total
The local policies, network access: Let Everyone permissions apply to anonymous users (applied automatically by Windows)
NFS Sharing - ANSI encoding, allow anonymous access - UID = - 2, GID = - 2
NFS share permissions - allow all THE MACHINES read-write, ROOT access
The data store adds OK and appears in the list of data on vClient stores and can be viewed or directly via a Console logon.
The problem is that the data store is read only! Nothing from ESX can write, delete, or update a folder or a file on the data store. Access Windows NTFS is very well.
This means also you cannot boot from say a CD ISO on the data store, or create a VMDK guest virtual computer on the data store.
Given that I had zero problems with this in the past, when with the help of ESX3 or ESX4 connect to Windows Server 2003R2 or several NAS with NFS devices supported I think Windows 2008R2?
Has anyone seen anything like this before?
I use it currently; It works fine, but it's the way that you configure the windows nfs export without having to use a mapping service (see images)
Also, be aware that Windows System Cache is a bit less efficient than linux (but that's ok) you just need to add additional RAM at least 16 GB for the data store busy.
I also recommend refining delays in setting up nfs for each virtual machine host:
NFS. HeartbeatFrequency = 10
NFS. HeartbeatFailures = 5
And on windows change the properties of the NFS server in the locking tab
customer collection lock wait 5
-
Excerpts from run.log
*****************
2014-09-10 13:21:51 UTC...
2014-09-10 13:21:51 UTC [VMSTAF] [0] INFO: VM: [PID 13775] command (perl /iozone - duration.pl d 3600 f - 128 m - r 1024 k t /tmp/iozone.dat 1 > /iozone.2.log 2 > & 1) on 192.168.1.25 is completed.
2014 09/11 01:36:16 UTC Test script failed to 300 s to/opt/vmware/VTAF/Certification/Workbench/VTAFShim 279, < STDIN > line 309 cleaning.
2014 09/11 01:36:16 UTC interrupted Test
One thing noticed here is that there is no logs generated during about 12 hours between the completion of test on one of the GOS and Test Script failed.
All of the recommendations will help...
Do not wait more than 2 hours if you think that the test case is stuck. Clean and restart the test.
-
Storage of P2V of NFS mounted in a virtual machine
Hello
I'm looking for the best method to achieve the following objectives:
We currently have a Server 2008 R2 Virtual Machine with its operating system disk in the virtual environment (vSphere 5, ESXi 5 Update1). In this machine are two additional disks provided by our EMC SAN using Microsoft iSCSI Initiator in the virtual machine. This our main file server.
We would like to take advantage of the software virtual backup we use (Veeam 6.1) so would to move these disks in the virtual environment as VMDK.
I know that potentially we could create two new VMDK and synchronize the data across, but is there a way to clone those to preserve all the permissions, etc.
Any advice would be appreciated!
Thank you
DanIt appears so, but I do not use Veeam Backup and replication, so I can't tell you for sure.
-
How to backup without identifier (NFS vs SMB/AFP)
Hello
you just bought a ReadyNAS 312 but do not yet know about some basic principles.
I want to use mainly as a backup destination, but also as the main source for backups Crashplan. In detail I will not install Crashplan headlessly on the NAS (totally not supported) but I prefer to install on my computer, which is then connected to the NAS through my line (just check, taken Crashplan supported files attached nas backups).
What I don't understand is how can automatically make sure Crashplan can see the files at all times... I think that this should be easy with NFS and Linux computers. I just need to configure actions to be mounted automatically at each start-up, right?
What, Mac computers, is even achievable with SMB/AFP so that I don't connect all the time in the Finder.
I know that these sounds of questions a bit silly, but it's the first time I have use a SIN 'serious '.
To be honest: sooner or later we will return only Linux boxes in the House, but at this point we can not yet get rid an Apple computer.
Thanks a lot for any suggestion
netghiro wrote:
My initial confusion is related to how to automatically assemble the network shares on my system (the a CrashPlan possibly running) to make them available to the local backup software (which will take care to copy the files to the NAS)... Sorry for the bad wording or maybe you misunderstand something.
http://support.code42.com/CrashPlan/latest/backup/Mounting_Networked_Storage_Or_NAS_Devices_For_Back... answers for OSX and links to a similar article for Windows. I suspect that for Linux, you simply create a NFS mount point suitable, but I have not tried. In any event, you don't really need to go with headless on the NAS installation, if it presents the advantage to avoid the need for a separate device.
netghiro wrote:
I think that when I'm on Linux and using NFS shares that will be easily magically mount every time you start (I still wonder if the username/password on the NAS must the same on my computer and to be honest, I was wondering the same for the two AFP/SMB!)
With Windows, the name of user and password may be different. Configure you in the Windows Credential Manager. I'm not a Mac guy, but I think that Keychain Apple does the same for network with OS x resources. I think that with Linux, you need to explicitly use the opening of session and password in the mount command (I don't think it's the equivalent of the Windows or Apple Keychain credentials manager).
netghiro wrote:
I have already subscribed to a (computer) personal unlimited plan, I can envision the family upgraded but happy enough to save a little money as well. I'll try to keep a couple of good games of local backup in the first place.Which is what I did. I have an unlimited plan for single computer for the NAS. I use Acronis image to back up the disks of PC 'C' on the NAS and also to use the SIN as a consolidated storage. I have 3 copies local of all (the primary account as a NAS) use and maintain CrashPlan as expensive disaster.
My pro - 6 has 2 GB of ram, which worked well for a few years, but eventually unlimited retention of deleted files caused central CrashPlan CrashPlan to run out of memory. This is what prompted the request for assistance to Crashplan. The solution was to the retention value of the files deleted in 6 months and then consolidate archive. That took a while (Crashplan was not wise to this point). I've also increased the memory in the pro and CrashPlan defined to use the size of the memory possible max (which, on the pro, is about 3.5 GB with a 32-bit JVM).
It has been working fine since that date (end of June). In any case, it's a good idea to lower the setting for the retention of deleted files. I also put deduplication to "minimal".
-
Hello
I intend to set up a node 2 CARS. However, I have only the internal drive to use and no SIN /SAN. So I intend to share some of the ability of alternatives I have on the internal nodes 1 & 2 discs and make shared storage using NFS. However I have a few questions about it.
Q1. Since the NFS mount will indeed be file system media, this means that for the storage for the Cluster option and the RAC database that I can only use option 'shared file system' and not 'ASM '?
I think what is confusing me is that I think that I read that the ASM disks can be complete physical disks, partitions of a disk, SAN LUNS or NFS?
If this is the case then it suggests that I can create an ASM disk based on a NFS mount?
I don't know how to do this - any help / advice would be appreciated because ideally, I would have preferred to create candidate ASM disks based on the shared NFS disk
Thank you
Jim
You can use NFS as your clustered file system, but if you want to use ASM we top of the page so it no problem: dd to create large files on the file system shared, for example,
If DD \u003d/dev/zero of = / mnt/nfs/asmdisk1 bs = 1048576 count = 2048
and set your asm_diskstring parameter to point to them (or specify them as devices to be used at installation time). It is very easy to set up and is fully supported by Oracle of the uncle. I recorded a few demos that include this a year or two before, Oracle ASM free tutorial
--
John Watson
Oracle Certified Master s/n
Maybe you are looking for
-
start of the tabs will not define
I'm new to Firefox (Google Chrome kept showing adware), I managed to import all of my settings, but I can't just not Firefox to start with a set of predefined tabs. I opened the option and created a folder of bookmarks and choose this folder in the "
-
I can not instalar plotter hp desingjet 42 500 not windows 7
-
My windows XP user name change?
Just go on windows, Now, for some unknown reason, he wants to ask me a username and password. I know how to change my password, but I don't know how to reset my username. I have not used the username for so long I forgot.
-
HP 240 G3 laptop: HP 240 mobile G3 Upgrade RAM help
Hello! I just bought a HP 240 G3 laptop and want to upgrade RAM 8 GB. Here are the specs of the laptop of HP support assistant: Model: HP 240 G3 Notebook PC Product no.: M1V30PA #ACJ Processor: Intel (r) Pentium (r) CPU N3540 2.16 GHz OS: Win 8.1 Cur
-
double set video card in place with the precision dell T7400 workstation
Hello I would like to know what the correct installation/configuration program is to have 2 GPUs running in the system. I am running windows 7 64 bit. I have a geforce gtx 295 running in the system and I recently installed my old quadro fx 1700, turn