Math, asked by kaurhasneet4939, 1 year ago

2 bulbs of 100w and 60w connected with 220v calculate the crrent

Answers

Answered by debtwenty12pe7hvl
0

Dear Friends, you did not mention it is series or parallel circuit  im giving u the both


                      IN SERIES CIRCUIT

1. When bulb are connected in series "CURRENT" is same , While "VOLTAGE" will be "DIVIDED" Across Each Circuit According to the Position of the circuit connected (As in Increase order)

2. And Voltage Is directly proportional to Light. If voltage is high than it will glow brightly and if voltage is low it will dim.

3. Now Consider V = 220v,

R1 = 100w

R2 = 60w

So P = V^2/r therefor R1 = (220)^2/100 = 484ohm

R2 = (220)^2/60 =806.7ohm

So According to the OHM's Law V = IR

Resistance Directly proportional to Voltage.

so Voltage Directly proportional to Light.

Hence 40 watt bulb will glow more Brightly rather than 100w.

   current in the circuit =V/total R  =220/[484ohm+806.7ohm]

    =220/1290.7  = 0.17 Amp ANS


          IN PARALLEL CIRCUIT

Total supply voltage = 220v

   Rated power of first bulb = 100 watt

   So, current through it, I=p/v= 100/220= 0.45 amps

   Resistance offered by first bulb =220/.45= 488 ohms

   Rated power of second bulb =60 watts

   Voltage is same, so current through is I=p/v= 60/220= 0.27 amps

   Resistance of second bulb = 220/.27=814 ohms

  So total current in the circuit is  [ 0.27 amps+0.45 amps]

                                         =0.72 amps    ANS

   


Similar questions