Not sure it makes sense thermodynamically to deal with waste heat that way (if you are transmitting “waste heat” in a narrow beam, you are basically just transmitting energy in a narrow beam, and so it’s not waste heat anymore—you can get useful work out of it).
edit: I suppose the question is: what % of the star’s outgoing energy can we harness in principle, such that waste heat is hard to tell apart from background, and we completely hide that the star is there. For example, in the limit, if you just used a little of the star’s energy to redirect all the rest into a black hole, will the waste heat just from the redirection effort be detectable? If so, we can’t hide a star, the best we can do is not use too much energy, so the star looks like a normal star with no life in the system (but evil aliens can still come and check it out, since they know a star is there). If not, maybe we can harness some bigger % on the way into the black hole. If so, what % is physically possible? I don’t know.
I feel like physicists already worked out that you can’t hide stars, but I don’t know the literature.
I suppose the question is: what % of the star’s outgoing energy can we harness in principle, such that waste heat is hard to tell apart from background
You can never have the temperature of outgoing radiation indistinguishable from the cosmic background, since energy is being generated by the star and in equilibrium more energy must leave than enters from the background.
The CMB reads as ~2.76k. Let’s say you wanted to radiate the entire energy output of a star at 3.76 k. That means the flux out equals the star’s flux plus the CMB flux in. For a star like the sun, the surface area of material required comes to a sphere a fifth of a light year (13000 AUs) in radius to dissipate a solar luminosity of energy (divide quantity of radiator material by the fraction of the solar luminosity you want to use, but keep in mind having the radiators closer than a 5th of a light year would probably be pointless since they’d be heated up by the sun) (also keep in mind such an object would look as large as the full moon 44 light years away and as wide as Jupiter in our sky 2,000 light years away). For ten kelvin, a sphere 0.025 light years or 1560 AUs in radius. For 50 kelvin, 62 AUs (twice as large as Neptune’s orbit).
Of course, there’s also the starlight flux of all the other nearby stars, which makes this worse for very low temperatures.
(Calculations done using energy out = energy in from CMB + solar flux, and the definition of blackbody radiation)
EDIT: I should go over some astronomy papers and figure out what amounts of material typically produce observable infrared excesses.
Not sure it makes sense thermodynamically to deal with waste heat that way (if you are transmitting “waste heat” in a narrow beam, you are basically just transmitting energy in a narrow beam, and so it’s not waste heat anymore—you can get useful work out of it).
edit: I suppose the question is: what % of the star’s outgoing energy can we harness in principle, such that waste heat is hard to tell apart from background, and we completely hide that the star is there. For example, in the limit, if you just used a little of the star’s energy to redirect all the rest into a black hole, will the waste heat just from the redirection effort be detectable? If so, we can’t hide a star, the best we can do is not use too much energy, so the star looks like a normal star with no life in the system (but evil aliens can still come and check it out, since they know a star is there). If not, maybe we can harness some bigger % on the way into the black hole. If so, what % is physically possible? I don’t know.
I feel like physicists already worked out that you can’t hide stars, but I don’t know the literature.
You can never have the temperature of outgoing radiation indistinguishable from the cosmic background, since energy is being generated by the star and in equilibrium more energy must leave than enters from the background.
The CMB reads as ~2.76k. Let’s say you wanted to radiate the entire energy output of a star at 3.76 k. That means the flux out equals the star’s flux plus the CMB flux in. For a star like the sun, the surface area of material required comes to a sphere a fifth of a light year (13000 AUs) in radius to dissipate a solar luminosity of energy (divide quantity of radiator material by the fraction of the solar luminosity you want to use, but keep in mind having the radiators closer than a 5th of a light year would probably be pointless since they’d be heated up by the sun) (also keep in mind such an object would look as large as the full moon 44 light years away and as wide as Jupiter in our sky 2,000 light years away). For ten kelvin, a sphere 0.025 light years or 1560 AUs in radius. For 50 kelvin, 62 AUs (twice as large as Neptune’s orbit).
Of course, there’s also the starlight flux of all the other nearby stars, which makes this worse for very low temperatures.
(Calculations done using energy out = energy in from CMB + solar flux, and the definition of blackbody radiation)
EDIT: I should go over some astronomy papers and figure out what amounts of material typically produce observable infrared excesses.
What does it mean to hide a star? Would it not be ‘visible’ by having gravity?
That explains dark matter — vast alien civilisations that leak nothing but gravity. And the microwave background.
No it doesn’t. Microwave background intensity is uncorrelated with imputed dark matter density in a given direction.